Home » Proposed System for Plagiarism Detection

Proposed System for Plagiarism Detection

Chapter 3

Save Time On Research and Writing
Hire a Pro to Write You a 100% Plagiarism-Free Paper.
Get My Paper

The Proposed System

  1. Introduction

This chapter introduces ZPLAG as proposed system, and its most important design issues are explained in details.

It is very easy for the student to find the documents and magazines using advanced search engines, so the problem of electronic thefts is no longer local or regional, but has become a global problem occurring in many areas.

Due to the Hugging of information, and correlation networks, the discovery of electronic thefts is a difficult task, and the discovery of the thefts started in the Arabic language and the most difficult task no doubt.

Save Time On Research and Writing
Hire a Pro to Write You a 100% Plagiarism-Free Paper.
Get My Paper

And in light of the growing e-learning systems in the Arab countries, this requires special techniques to detect thefts electronic written in Arabic. And although it could use some search engines like Google, it is very difficult to copy and paste the sentences in the search engines to find these thefts.

For this reason, it must be develop a good tool for the discovery of electronic thefts written Arabic language to protect e-learning systems, and to facilitate and accelerate the learning process, where it can automatically detect electronic thefts automatically by this tool.

This thesis shows, ZPLAG, a system that works on the Internet to enable specialists to detect thefts of electronic texts in Arabic so it can be integrated with e-learning systems to ensure the safety of students and research papers and scientific theses of electronic thefts.

The thesis also describes the major components of this system, including stage outfitted, and in the end we will establish an experimental system on a set of documents and Arabic texts and compared the results obtained with some of the existing systems, particularly TurnItIn.

The chapter is organized as follow; Section 3.2 presents an overview of the Arabic E-Learning, Section 3.3 presents and explains the General Overview of the Proposed System, Section 3.4 explains in details the system architecture of the proposed system ZPLAG. Section 3.5 gives a summery for this chapter.

  1. General Overview of the Proposed System

The proposed system consists of three different phases namely; (1) Preparation phase, (2) Processing phase, and (3) Similarity detection phase. Figure 3.1 depicts the phases of the proposed system.

Figure 3.1 Proposed system phases

Preparation Phases: this phase is responsible for collecting and prepares the documents for the next phase. It consists of five modules: text editor module, check language module, check spelling module, check grammar module, and Sentences analysis module.

Text editor module allows the user to input a text or upload a text file in document format, these files can be processed in the next phase.

The check language module is responsible for checking the input file written language, If it is an Arabic language then use Arabic process, or English language then use English process.

The check spelling module use to check the words are written correct or there is some misspelling.

This phase consists of three modules explained as follows:

  1. Tokenization: break up the input text as some token .
  2. SWR: remove the common words that appear in the text but carry little meaning.
  3. Rooting: is the process of removing: (prefixes, infixes, or/and suffixes) from words to get the roots or stems of this word
  4. Replacement of Synonym: words are converted to their synonyms.

Similarity detection Phases: It is consists of three modules Fingerprinting, documents representation and similarity detection, this phase discussed as follows: To calculate fingerprints of any document, first cut up the text into small pieces called chunks, the chunking method that responsible for cutting up the text will be determined [12]. A unit of chunk could be a sentence or a word. In case of chunking using sentences called sentence-based, the document can be cutted into small chunks based on ‘C’ parameter. For example, a document containing sentences ds1 ds2 ds3 ds4 ds5, if C=3 then the calculated chunks will be ds1 ds2 ds3, ds2 ds3 ds4, ds3 ds4 ds5. For example, a document containing words dw1 dw2 dw3 dw4 dw5, if C=3 then the calculated chunks will be dw1 dw2 dw3, dw2 dw3 dw4, dw3 dw4 dw5. The chunking using Word gives higher precision in similarity detection than the chunking sentence.

  1. The Architecture pf Proposed System

The following properties should be satisfied by any system detecting plagiarism in natural language:

  • Insensitivity to small matches.
  • Insensitivity to punctuation, capitalization, etc.
  • Insensitivity to permutations of the document content.

The system main architecture of ZPLAG is illustrated in Figur1.

  • Preparation: text editor, check language, check spelling, and check grammar.
  • Preprocess: synonym replacement, tokenization, rooting, and stop-word removal.
  • Fingerprinting: the use of n-gram, where the user choses the parameter n.
  • Document representation: for each document, create a document tree structure that describes its internal representation.
  • Selection of a similarity: use of a similarity metric to find the longest match of two hash strings.

As mentioned in the previous section, the system architecture breakdown contains three main phases. Each phase will be composed to a set of modules in terms of system functionality. The following section contains the description of each phase and its modules in details.

3.4.1 The Preparation Phase

The main task of this phase is to prepare the data for the next phase. It consists of text editor module, check language module, check spelling module and check grammars module.

3.4.1.1. Text editor Module

Figure 3.2, illustrates text editor module. The users of the text editor module are faculty members and students, where the users need a text area to upload their files, so the brows helps for file path to make it easy for the users, After that check file format is very important , because the service upload files with doc or docx format, then after the user upload the file , the text editor module save the file in the database.

Figure 3.2 text editor module

 

3.4.1.2 Check Language Module

The raw text of the document is treated separately as well. In order to extract terms from text, classic Natural Language Processing (NLP) techniques are applied as. Figure 3.3 illustrates Check Language module and its functions: from the system database, whereas all the files are stored, the check language module bring the file and read it, then check for language either Arabic , English or combo (both Arabic and English), After that mark the document with its written language and save the file again in the system database.

Figure 3.3 check language module

3.4.1.3 Check Spelling Module

Figure 3.4 illustrates Check spelling module and its functions: after bringing the document from the system database, whereas all the files are stored, the check spelling module read the file, and use the web spelling checker, then the check spelling module make all the possible replacements for the words in false spelling check , After that save the file again in the system database.

Figure 3.4 check spelling module

3.4.1.4 Check Grammars Module

For English documents, Figure 3.5 illustrates Check grammar module and its functions: after bringing the document from the system database, whereas all the files are stored, the check grammar module read the file, and use the web grammar checker, After that the check grammar module mark the sentences with the suitable grammar mark and save the file again in the system database.

 

Figure 3.5 check grammar module

3.4.2 The processing Phase

3.4.2.1 The Tokenization Module

In the Tokenization module : after bringing the document from the system database, whereas all the files are stored, the Tokenization module read the file, and brake down the file into paragraphs, after that brake down the paragraphs into sentences, then brake down the sentence into words. After that save the file again in the system database.

3.4.2.2 The Stop Words Removal and Rooting Module

The raw text of the document is treated separately as well. In order to extract terms from text, classic Natural Language Processing (NLP) techniques are applied as. Figure 3.6 illustrates Stop Words Removal and rooting module and its functions:

Figure 3.6: SWR and Rooting module

SWR: Common stop words in English include: a, an, the, in, of, on, are, be, if, into, which etc. Whereas stop words in Arabic include: من , إلى , عن , على , في …etc. These words do not provide a significant meaning to the documents . Therefore, they should be removed in order to reduce ‘noise’ and to reduce the computation time.

Word Stemming: it will be changed into the word’s basic form.

3.4.2.3 Replacement of Synonym

Replacement of Synonym: It may help to detect advanced forms of hidden plagiarism. The first synonym in the list of synonyms of a given word is considered as the most frequent one.

3.4.3 The Similarity Detection Phase

3.4.3.1 The Fingerprinting Module

It is consists of three modules Fingerprinting, documents representation and similarity detection, this phase discussed as follows: To calculate fingerprints of any document, first cut up the text into small pieces called chunks, the chunking method that responsible for cutting up the text will be determined [12]. A unit of chunk could be a sentence or a word. In case of chunking using sentences called sentence-based, the document can be cutted into small chunks based on ‘C’ parameter. For example, a document containing sentences ds1 ds2 ds3 ds4 ds5, if C=3 then the calculated chunks will be ds1 ds2 ds3, ds2 ds3 ds4, ds3 ds4 ds5. In case of chunking using word called a word-based chunking, the document is cutted into small chunks based on ‘C’ parameter. For example, a document containing words dw1 dw2 dw3 dw4 dw5, if C=3 then the calculated chunks will be dw1 dw2 dw3, dw2 dw3 dw4, dw3 dw4 dw5. The chunking using Word gives higher precision in similarity detection than the chunking sentence. ZPLAG is based on a word-based chunking method: in every sentence of a document, words are first chunked and then use a hash function for hashing.

3.4.3.2 The Document Representation Module

Document representation: for each document, create a document tree structure that describes its internal representation.

3.4.3.3 The Similarity Detection Module

A tree representation is created for each document to describe its logical structure. The root represents the document itself, the second level represents the paragraphs, and the leaf nodes contain the sentences.

  1. Summary

Being a growing problem, The electronic thefts is generally known as “plagiarism” and “dishonesty academic” and they constitute a growing phenomenon, It should be known that way to prevent its spread and preserve the ethical principles that control the academic environments, with easy access to information on the World Wide Web and the large number of digital libraries, electronic thefts have become one of the most important issues that plague universities and scientific centers and research.

This chapter presented in detailed description of the proposed system for plagiarism detection in electronic resources and its phases and its functions.

Place your order
(550 words)

Approximate price: $22

Calculate the price of your order

550 words
We'll send you the first draft for approval by September 11, 2018 at 10:52 AM
Total price:
$26
The price is based on these factors:
Academic level
Number of pages
Urgency
Basic features
  • Free title page and bibliography
  • Unlimited revisions
  • Plagiarism-free guarantee
  • Money-back guarantee
  • 24/7 support
On-demand options
  • Writer’s samples
  • Part-by-part delivery
  • Overnight delivery
  • Copies of used sources
  • Expert Proofreading
Paper format
  • 275 words per page
  • 12 pt Arial/Times New Roman
  • Double line spacing
  • Any citation style (APA, MLA, Chicago/Turabian, Harvard)

Our guarantees

Delivering a high-quality product at a reasonable price is not enough anymore.
That’s why we have developed 5 beneficial guarantees that will make your experience with our service enjoyable, easy, and safe.

Money-back guarantee

You have to be 100% sure of the quality of your product to give a money-back guarantee. This describes us perfectly. Make sure that this guarantee is totally transparent.

Read more

Zero-plagiarism guarantee

Each paper is composed from scratch, according to your instructions. It is then checked by our plagiarism-detection software. There is no gap where plagiarism could squeeze in.

Read more

Free-revision policy

Thanks to our free revisions, there is no way for you to be unsatisfied. We will work on your paper until you are completely happy with the result.

Read more

Privacy policy

Your email is safe, as we store it according to international data protection rules. Your bank details are secure, as we use only reliable payment systems.

Read more

Fair-cooperation guarantee

By sending us your money, you buy the service we provide. Check out our terms and conditions if you prefer business talks to be laid out in official language.

Read more
Live Chat+1 763 309 4299EmailWhatsApp

We Can Handle your Online Class from as low as$100 per week