The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
The World Wide Web has become a huge repository of data of interest for a variety of application domains. However, the same features that have made the Web so useful and popular also impose important restrictions on the way the data it contains can be manipulated. Particularly, in the traditional Web scenario, there is an inherent difficulty in gaining access to data that is implicitly present in...
of web resources. Thus finding the information related to specific topic or a keyword from largely available web resources enables promising opportunities for information discovery. In this paper we propose a system that finds the links related to specific keyword and then it performs the in-site searching to get the
systems, and they are publicly available through the online repositories of popular CMTs. However, locating and retrieving desired operation services is challenging since keyword-and tag-based search provided by a repository is with low precision. In this paper, we implement a hierarchical categorization approach based
overloaded sites for a short piece of information of their interest. The crawler developed in the system gathers web page information which is processed using Natural Language Processing and Procedure programming for a specific keyword. The system returns precise short string answers or list to natural language questions
to determine the forms' relevance to the domain. In this work scientific research publications domain has been considered. Experimental results show that proposed approach is better as compared to keyword based crawlers in terms of both relevancy and completeness.
keyword queries through a standard search engine and receive up-to-date database information. The system was then tested to determine if it could return results that were similar to those submitted using SQL. We also looked at whether a standard search engine such as Google could actually index the database content
Search engine is the most important tool to discover the any information in World Wide Web. In a row with the terrific growth numbers of the Web, traditional search engine nowadays is not appropriate anymore to be used. Searched by keyword and do not understand polysemy and synonymy are some reasons the traditional
. The first part of this paper explores illicit and criminal content identified by prominent researchers in the dark web. We previously developed a web crawler that automatically searched websites on the internet based on pre-defined keywords and followed the hyperlinks in order to create a map of the network. This crawler
system. In this paper, an OntoCrawlerII based on the ontology-supported technique for webpage searching was proposed, in which scholars just need to enter some keywords and the system supported by domain ontology actively would provide comparison and verification for those keywords so as to up-rise precision rate and recall
keywords, which lack the semantic data. In this paper the semantic based information retrieval methodology is proposed to get data from the web archives in a specific domain by gathering the domain relevant information with web crawler. By utilizing ontology and semantic information matched with a given user's query is used
uses predefined keywords segmented in the web page text to calculate the shortest distance. The face date of science and technical experts downloaded are deposited in local folder, and then the absolute path got by means of folder traversal is stored in the MySQL database system which can achieve the query and statistics
As an ever-increasing amount of information on the Web today is available through search interfaces, users have to key in a set of keywords in order to access the pages from certain Web sites, which are often referred to as the hidden Web or the deep Web. Since there is no static links to the hidden Web pages, search
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.