The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
distances in a multidimensional scaling space. In this study, we introduce an example of a 3-D multimedia space using the Associated Keyword Space (ASKS) and demonstrate similarity relationships between various sources of data in this space.
This paper presents a new keyword extraction algorithm for Chinese news Web pages using lexical chains and word co-occurrence combined with frequency features, cohesion features, and corelation features. A lexical chain is an external performance consistency by semantically related words of a text, and is the
A keyword choice and analysis approach in SEO is studied to deal with the issues such as low efficiency, poor reliability and instability optimization in artificial SEO processing in this paper. A keyword expansion method is proposed by reversing search engine's related search keywords to meet user's requirements
Nowadays the existing search engines are always lack of the consideration of personalization. They display the same search results for different users despite their differences in interesting and purpose. In order to solve this problem, this paper introduces a new method of using keyword query series to express the
keywords from the Web pages. The system first identifies the section of the Web page that contains the multimedia file to be extracted and then extracts it by using clustering techniques and other tools of statistical origin. Experimental results on real-world image sharing Web sites are presented and discussed in this paper
Web pages for search engine. First we describe a scheme based on semantic keywords combined with sentence overlapping, and then show an implemented prototype, with the experimental results that suggest the prototype work well under a proper setting.
In order to improve searching results of Web pages and enhancing Web crawling operation, the Web page clustering based on searching keywords is proposed in this paper, which firstly employed matching degree between Web pages and searching keywords to decide the sequence of showing pages of searching results. Then
This paper presents an attempt to show the efficiency of some search engines in dealing with Arabic keywords. This can be achieved by comparing the number of retrieved pages, retrieving time, and stability (in both the number of retrieved pages and the order for each retrieved page) for each one of the selected 20
It's important to eliminate noisy data for information extraction on the deep web. In this paper, we propose a new approach called ENDW(Eliminating Noisy Data in Web pages) based on query keywords and DOM tools to eliminate noisy data. Query keywords submitted to backend databases always appear in deep web pages. The
The traditional layout of news websites, the combination of classified hierarchical browsing, headline recommendation and keyword-based search, has been used for many years. The keyword-based search is considered to be the most powerful tool for news browsing and retrieval. Unfortunately, the keyword-based query
Search engines are one of the most powerful tools in the Web world today for data retrieval and exploration. Most search engines identify the key word in the sentence or phrase or list of words given by the user and starts mining the Web for the occurrence of keyword in the Web pages. Quite often searching for the key
Nowadays, web has become widespread in terms of availability of contents related to every field. Also a large repository of web contents is turned up as a most challenging tool for searching and retrieving information. For scientists and researchers, resource or content searching has been very important. Todays, market is full of variant search tools over web having discrepancy in terms of working...
-processing of Web search results have been extensively studied to help user effectively obtain useful information. This paper has basically three parts. First part is the review study on how the keyword is expanded through truncation or wildcards (which is a little known feature but one of the most powerful one) by using
needed to search and find relevant information. For tabular structures embedded in HTML documents, typical keyword or link-analysis based search fails. The next phase envisioned for the WWW is automatic ad-hoc interaction between intelligent agents, web services, databases and semantic web enabled applications. A large
The World Wide Web has become a huge repository of data of interest for a variety of application domains. However, the same features that have made the Web so useful and popular also impose important restrictions on the way the data it contains can be manipulated. Particularly, in the traditional Web scenario, there is an inherent difficulty in gaining access to data that is implicitly present in...
the crawled Web pages in to repositories. At first, the keywords are extracted from the crawled pages and the similarity score between two pages is calculated based on the extracted keywords. The documents having similarity scores greater than a threshold value are considered as near duplicates. The detection has
queries, reverse queries, Webpage title and keyword phrases are combined with the cluster centers to attain high-quality expansion terms for new queries. We also propose a new terminology extraction method through Baidu Baike. It can identify and extract the terminology phrase based on the manual edited dictionary online.
semantic net which can be applied to build personalized search engine and tested with single query keyword and multi ones by three different calculating policies. The test results show that it can affect the sort of pages. The personalized search based on vocabulary semantic net improves the quality of search results greatly.
The motivation behind sub-topic or topic specific keyword discovery through Web pages is helping a user, who is insufficient in knowledge and experience about a topic, to find important concepts without much effort. Intuitively, a Web user would start searching the Web via querying search engines, visiting some pages
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.