Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 14-Day Trial for You or Your Team.

Learn More →

A structural, content‐similarity measure for detecting spam documents on the web

A structural, content‐similarity measure for detecting spam documents on the web Purpose – The web provides its users with abundant information. Unfortunately, when a web search is performed, both users and search engines must deal with an annoying problem: the presence of spam documents that are ranked among legitimate ones. The mixed results downgrade the performance of search engines and frustrate users who are required to filter out useless information. To improve the quality of web searches, the number of spam documents on the web must be reduced, if they cannot be eradicated entirely. This paper aims to present a novel approach for identifying spam web documents, which have mismatched titles and bodies and/or low percentage of hidden content in markup data structure. Design/methodology/approach – The paper shows that by considering the degree of similarity among the words in the title and body of a web docuemnt D , which is computed by using their word‐correlation factors; using the percentage of hidden context in the markup data structure within D ; and/or considering the bigram or trigram phase‐similarity values of D , it is possible to determine whether D is spam with high accuracy Findings – By considering the content and markup of web documents, this paper develops a spam‐detection tool that is: reliable, since we can accurately detect 84.5 percent of spam/legitimate web documents; and computational inexpensive, since the word‐correlation factors used for content analysis are pre‐computed. Research limitations/implications – Since the bigram‐correlation values employed in the spam‐detection approach are computed by using the unigram‐correlation factors, it imposes additional computational time during the spam‐detection process and could generate higher number of misclassified spam web documents. Originality/value – The paper verifies that the spam‐detection approach outperforms existing anti‐spam methods by at least 3 percent in terms of F ‐measure. http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png International Journal of Web Information Systems Emerald Publishing

A structural, content‐similarity measure for detecting spam documents on the web

Loading next page...
 
/lp/emerald-publishing/a-structural-content-similarity-measure-for-detecting-spam-documents-PUNJCaLJCU

References (37)

Publisher
Emerald Publishing
Copyright
Copyright © 2009 Emerald Group Publishing Limited. All rights reserved.
ISSN
1744-0084
DOI
10.1108/17440080911006207
Publisher site
See Article on Publisher Site

Abstract

Purpose – The web provides its users with abundant information. Unfortunately, when a web search is performed, both users and search engines must deal with an annoying problem: the presence of spam documents that are ranked among legitimate ones. The mixed results downgrade the performance of search engines and frustrate users who are required to filter out useless information. To improve the quality of web searches, the number of spam documents on the web must be reduced, if they cannot be eradicated entirely. This paper aims to present a novel approach for identifying spam web documents, which have mismatched titles and bodies and/or low percentage of hidden content in markup data structure. Design/methodology/approach – The paper shows that by considering the degree of similarity among the words in the title and body of a web docuemnt D , which is computed by using their word‐correlation factors; using the percentage of hidden context in the markup data structure within D ; and/or considering the bigram or trigram phase‐similarity values of D , it is possible to determine whether D is spam with high accuracy Findings – By considering the content and markup of web documents, this paper develops a spam‐detection tool that is: reliable, since we can accurately detect 84.5 percent of spam/legitimate web documents; and computational inexpensive, since the word‐correlation factors used for content analysis are pre‐computed. Research limitations/implications – Since the bigram‐correlation values employed in the spam‐detection approach are computed by using the unigram‐correlation factors, it imposes additional computational time during the spam‐detection process and could generate higher number of misclassified spam web documents. Originality/value – The paper verifies that the spam‐detection approach outperforms existing anti‐spam methods by at least 3 percent in terms of F ‐measure.

Journal

International Journal of Web Information SystemsEmerald Publishing

Published: Nov 20, 2009

Keywords: Accuracy; Error analysis; Data structures

There are no references for this article.