Web crawler thesis duplicate checking method
A web crawler and paper technology, applied in the field of web crawler, can solve the problems of large workload of checking, large amount of articles, low accuracy rate, etc.
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment Construction
[0017] A method for checking plagiarism of web crawler papers, the specific steps are as follows:
[0018] The first step is to upload the papers that need to be checked for plagiarism. The uploaded papers that need to be checked for plagiarism can be in various formats such as WORD and PDF.
[0019] The second step is to extract the keyword groups in the paper. The keyword groups are phrases, paragraphs and symbols that appear frequently in the article. There is also an input window here, and I can perform manual operations.
[0020] In the third step, the server searches for article URLs related to the keywords of the uploaded articles, and the initial webpage information of the URLs of the relevant information articles is retrieved, and there will be no information window to display the information.
[0021] The fourth step is to grab articles similar to the uploaded articles in the paper information database.
[0022] The fifth step is to compare the captured article with...
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More 