The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
Archiving graph data over history is demanded in many applications, such as social network studies, collaborative projects, scientific graph databases, and bibliographies. Typically people are interested in querying temporal graphs. Existing keyword search approaches for graph-structured data are insufficient for
The study on XML keyword search gradually becomes the focus of information retrieval. Most previous XML keyword search algorithms are based on SLCA (smallest lowest common ancestor), but in the process of keyword search, we discover that some weakness or flaw exists in SLCA, it is summarized as follows: (1) the query
With the XML becomes a de-facto standard for exchanging and presenting information, the study on XML keyword search has become the focus of information retrieval. Several recent studies have finished the effective XML keyword search, but not all approach is effective in identifying return information, not all search
A lot of semantic information is lost due to keyword centric approach of information indexing. Web search should be based on `context' of the query and not only on the keywords in query. It is only possible when a context from a query as well as document is sensed and which requires a context based indexing approach
We introduce a new method for discovering latent topics in sets of objects, such as documents. Our method, which we call PARIS (for Principal Atoms Recognition In Sets), aims to detect principal sets of elements, representing latent topics in the data, that tend to appear frequently together. These latent topics, which we refer to as `atoms', are used as the basis for clustering, classification, collaborative...
terrain data block cache, and discusses two major items, the design of keyword and the solution of collision. The result of experiment suggests that this hash indexing algorithm has low collision and can retrieve cache data fleetly and efficiently.
Latent Dirichlet Allocation, which is a non-supervised learning method, can be used for topic detection, automatic text categorization, keyword extraction and so on. It only focuses on the text itself, not considering other external correlation properties. External association property refers to some structured
Domain Assets are the domain knowledge constructed according to the common requirements in the domain. In order to reuse the domain assets effectively, a domain assets search algorithm is proposed in this paper. Compared with the keyword search, this algorithm is based on semantic similarity, and the domain assets
special data record and new record model on fuzzy set are given. By calculating the membership of keyword, new fuzzy closeness functions are proposed to classify the information. Finally, examples prove that this algorithm can effectively and automatically classify input information of database, the accuracy and intelligence
always ignores relativity of the topic. These affect the topic discovery and topic trend. Therefore, combining with the keywords combination and Word2Vec model to strength expression of semantic information in topic clustering, this article sets weighted K-means algorithm for topic discovery. The results show our weighted K
In the processing of source retrieval in plagiarism detection, rationale for keywords extraction is to select only those phrases or words which maximize the chance of retrieving source documents matching the suspicious document. TF-IDF (term frequency-inverse document frequency), weighted TF-IDF (the weighted term
The study aimed at analyzing the keywords of the Macau Special Administrative Region's 2012 and 2013 annual policy addresses. The contribution of the study included the following two points. First, the study used the text mining method in order to explore the content of policy address. Second., the study applied the
and miss important emails from important people. This email management issue imposes an adverse effect on the productivity of email communication. Although many email clients today are equipped with tools to filter emails based on keywords, email addresses; most of these filters are static and are not updated
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.