The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
easy to bring the problem of topic excursion. Hits algorithm requires a number of pages as the basic-set for calculating and cannot be used in plain texts. This paper introduces a new algorithm: PK-TDC which makes use of the iterative idea of Hits. PK-TDC searches the authority pages and keywords on the topology of pages
Traditional Web search engines mostly adopt a keyword-based approach. When the keyword submitted by the user is ambiguous, search result usually consists of documents related to various meanings of the keyword, while the user is probably interested in only one of them. In this paper we attempt to provide a solution to
Users of search engines interact with the system using different size and type of queries. Current search engines perform well with keyword queries but are not for verbose queries which are too long, detailed, or are expressed in more words than are needed. The detection of verbose queries may help search engines to
(MWE) and they do not scale very well. This paper proposes a clustering and classification algorithm for semantic similarity using sample web pages. Further improvement is to analyze the short text for classification and labeling the short text according to the keyword and producing the result for the end user. This type
(MWE) and they do not scale very well. This paper proposes a clustering and classification algorithm for semantic similarity using sample web pages. Further improvement is to analyze the short text for classification and labeling the short text according to the keyword and producing the result for the end user. This type
There are huge numbers of valuable information resources resided on Invisible Web. However, it is hard to use for us. In this paper we propose a system called NewsReaper that is capable of making Invisible Web to be visible, especially the huge number of real-time information, which update frequently and are time-sensitive. NewsReaper makes use of information extraction, text classification, full...
underlying techniques used in these engines. These techniques are mostly based in the frequency of the keywords of the query in the HTML code. In addition, issues such as dealing with classifying the pages found for a query according to previous visits along with features needed to make intelligent decisions regarding the
small number of HTML input elements extracted from user input HTML forms and a few keywords. It utilizes pre-query technique and post-query technique in a hierarchical manner. Decision trees and multi layer artificial neural networks were used to obtain the classification rates over 91% to classify search forms and non
likely encountered a high ranking page that consists of nothing more than a bunch of query keywords. These pages detract both from the user experience and from the quality of the search engine. Search engine spam is a webpage that has been designed to artificially inflating its search engine ranking. Recently this search
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.