The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
User interaction with web sites generates a large amount of web access data stored in the web access logs. Those data can be used for e-commerce to conduct an evaluation of possessed website pages as one of the efforts to understand the desires of the user. Through classification techniques in web usage mining, we conducted an experiment to categorize a number of data obtained from the client log...
In the blogosphere, the amount of digital content is expanding and for search engines, new challenges have been imposed. Due to the changing information need, automatic methods are needed to support blog search users to filter information by different facets. In our work, we aim to support blog search with genre and facet information. Since we focus on the news genre, our approach is to classify blogs...
In Web database integration, crawling data pages is important for data extraction. The fact that data are contained by multiple result pages increases the difficulty of accessing data for integration. Thus, it is necessary to accurately and automatically crawl query result pages from Web database. To address this problem, we propose a novel approach based on URL classification to effectively identify...
Mutual information formula is improved by using the hyperlink factor in this paper. Introduction of hyperlink elements of web pages can improve the classification accuracy in feature selection method based on mutual information and correlation by experiment, especially for those of strong. So the improvement is effective in web page classification.
Text classification is one of the core applications in data mining due to the huge amount of not categorized digital data available. Training a text classifier generates a model that reflects the characteristics of the domain. However, if no training data is available, labeled data from a related but different domain might be exploited to perform cross-domain classification. In our work, we aim to...
This paper introduces an approach to classifying emails into phishing/non-phishing categories using the C5.0 algorithm which achieves very high precision and an ensemble of other classifiers that achieve high recall. The representation of instances used in this paper is very small consisting of only five features. Results of an evaluation of this system, using over 8,000 emails approximately half...
Emotion plays a significant role in human communications in our daily life. With progress in human-machine interface technology, recent research has placed more emphasis on the recognition of emotion reaction. Comparing to some other ideal experimental settings, blog posts online would be respond more to real-world events. And a huge resource of text-based emotion can be found from the World Wide...
In this paper we explore the task of mood classification for blog postings. We propose a novel approach that uses the hierarchy of possible moods to achieve better results than a standard machine learning approach. We also show that using sentiment orientation features improves the performance of classification. We used the Livejournal blog corpus as a dataset to train and evaluate our method.
Client honeypots are security devices designed to find servers that attack clients. High-interaction client honeypots (HICHPs) classify potentially malicious Web pages by driving a dedicated vulnerable Web browser to retrieve and classify these pages. Considering the size of the Internet, the ability to identify many malicious Web pages is a crucial task. HICHPs, however, present challenges: They...
We consider the problem of content extraction from online news Web pages. To explore to what extent the syntactic markup and the visual structure of a Web page facilitate the extraction of its content, we compare two state-of-the-art classifiers as first instantiations of a general framework that allows for proper model comparison. To this end, we introduce the publicly available NEWS600 corpus, a...
Nowadays, any employment and recruitment Web sites receive immense personal information and recruit information every day. But most information canpsilat be properly analyzed and canpsilat meet the recruit requirement. In fact, the recruiting units are looking for talents of both high and low levels talents. However, many talents information canpsilat be evaluated correctly so that the appliers lose...
In this paper, reclassification for the current classification through K-means would be implemented based on the feedback of Web usage mining in order to improve the accuracy of news recommendation and convergence of classification. It could extract most relative keywords and eliminate the disturbance of multi-vocal word in one category based on feedback of Web usage. The reclassification of news...
In this paper, an efficient text classification algorithm for repeating-text information on the e-commerce site can automatically classify and sort the similar string. This algorithm will greatly increase the efficiency and accuracy of audited information. All tests show that for the number of information between 100 and 1000 the algorithm is very efficient, and the 1000 text information(strings)...
Traditional automatic classifiers often conduct misclassifications. Folksonomy, a new manual classification scheme based on tagging efforts of users with freely chosen keywords can effective resolve this problem. Even though the scalability of folksonomy is much higher than the other manual classification schemes, the method cannot deal with tremendous number of items such as whole Weblog articles...
Two important factors which indirectly influence the Internet shoppers to make some online purchases are the visual layout and the presentation of web page. In this paper, we propose an approach of web page layout analysis in order to assess the design of e-commerce Web sites. Firstly, our proposed method segments each web page into five different blocks: top, left, center, right and bottom. We study...
Among a huge number of blogs on the internet, only some of them are considered to have great contents and worth to be explored. We call such kind of blogs cool blogs and attempt to identify them. To solve the cool blog identification problem, we consider three assumptions on cool blogs: (1) cool blogs tend to have definite topics, (2) cool blogs tend to have sufficient amount of blog entries, and...
Reprinting information among websites produces a great deal redundant WebPages. To improve search efficiency and user satisfaction, an algorithm to Detect near-Duplicate WebPages (DDW) is proposed. In the course of developing a near-duplicate detection system for a multi-billion page repository, we make two research contributions. First, we consider both syntactic and semantic information to present...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.