The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
Many important forms of data are stored digitally in XML format. Errors can occur in the textual content of the data in the fields of the XML. Fixing these errors manually is time-consuming and expensive, especially for large amounts of data. There is increasing interest in the research, development, and use of automated techniques for assisting with data cleaning. Electronic dictionaries are an important...
Conventional approaches to XML clustering by content and structure are generally affected by a limitation due to the adoption of the bag-of-word model for the representation of their textual contents. This choice may lead to consider structure-constrained textual items of separate XML documents as related, even though the actual meaning of such items in their respective contexts is different. To overcome...
Data in smart cities is commonly generated by a large variety of participants including institutional actors, equipment manufacturers, network operators, infrastructure providers, service providers, and end users. This data potentially undergoes several transformations such as aggregation and/or composition before finally being consumed. In this context of sharing data between diverse consumers, it...
Simulators are an essential tool for researching and developing new systems, architectures or protocols. They have played a key role in the history of the Internet and the Web. This paper highlights the view that the simulation-based research will continue to take part in the evolution of paradigms associated with the Internet and the Web of Things. In this work, we present a simulator whose ultimate...
In the Internet of Things (IoT) era, we need to face increased masses of cross-domain data stored in different formats (either relational, XML, JSON, textual) and data streams (produced by sensors), that can be highly or loosely structured and need to be integrated for analysis. Recently, many NoSQL systems (e.g., MongoDB, Cassandra, HBASE) have been born for coping the scalability issues of current...
State-of-the-art in duplicate detection in semi-structured data obtains significant improvement by exploiting the schema-related knowledge. Such schema-bound duplicate detection approaches, however, have severe limitations when dealing with multi-sourced, heterogeneous, high-velocity data streams. In this paper, we propose a novel context-aware duplicate detection system which is workload- and complexity-aware,...
Generating source code from a model, using a model-driven development techniques (MDD), rapidly increases application reliability and substantially accelerates time-to-market, thereby reducing application development costs. The generated source code does not have its functional implementation, so besides MDD, it is necessary to create some additional tools that will fulfill this gap. This paper describes...
Currently managing information overload has become a major challenge. How to manage all these data, presented in diverse formats and originating from heterogeneous sources? This paper presents a strategy to perform data fusion effectively. Our strategy deals with the problem of object identification in the context of the Command and Control of the Brazilian Defense Ministry using MIP Data Model from...
The popularity of the Internet and the rapid development of Web technology and its application make the XML technology developed quickly. More and more programmers store the data in XML form. With the large number of XML documents, applications need to continue to deal with XML data, access and update the XML data. Then XQuery is emerged. This paper introduces the XQuery language briefly, discusses...
Today, information systems rely largely on web services and XML as their payloads. Every information system also depends on checking and preserving integrity constraints. In the world of conceptual modeling, they can be expressed using Object Constraint Language (OCL) expressions over class diagrams. However, the expressions defined in the conceptual model of the system must be rewritten to the form...
Information sharing is a key requirement in Live, Virtual, and Constructive (LVC) simulation environments. Operational plans, orders, and requests from live, virtual, or constructive command and control systems or simulations need to be received by and operated on by receiving LVC systems. Situational reports from the LVC systems need to be received and interpreted or displayed by receiving LVC systems...
We propose a new approach to XML classification, that uses a particular rule-learning technique for the induction of interpretable classification models. These separate the individual classes of XML documents by looking at the presence within the XML documents themselves of certain features, that provide information on their content and structure. The devised approach induces classifiers with outperforming...
Tool support is essential for Application Engineering in Software Product Lines (SPL). Despite a myriad of existing tools, most lack adequate support for flexibility and adaptability, so that it is hard for them to be applied in different contexts, e.g., addressing variability in different artifacts. Addressing this issue requires exploring underlying commonality and adequately managing variability...
This paper discusses the modern approach of the implementation of Software as a Service (SaaS) for NASA as a way to reduce cost and increase efficiency. The Grid XML Datastore Framework is an extension of SaaS framework based on eXtensible Database technology (XDB). This is implemented to support the Innovative Partnership Program (IPP) for its Technology Transfer System (NTTS) project. NTTS supports...
The XML DOM (Document Object Model) provides a logical view of the in-memory structure. It represents the metadata that has a hierarchical treelike structure consisting of nodes. There are several benefits of implementing the XML DOM as a home management server. First of all, the process time decreases by three times in average sense than using general database. Secondly, it would make to meet with...
With XML becoming the de facto standard for representing and exchanging data on the Internet, the problem of validation of XML data when updated has attracted more and more attentions. The traditional brute-force validation processes the entire updated XML document from scratch, which is less efficient. This paper presents a method for XML incremental validation based on simplified XML element sequence...
In this paper, we address the problem of modeling the constraints of a mediated schema. We argue that, from the point of view of an application that processes the results of queries defined over the mediated schema, the constraints should be modeled as the greatest lower bound of the constraints of the export schemas, after appropriate translation to a common vocabulary. This assures that the application...
XML has become the de facto standard for representing and interchanging data in web-based applications. And XML view, a virtual window for specified users, has been widely applied. In practical system, users encounter the so-called view update problem when they need update source data through the view. For a long time, the view update problem is an open question in database community. With the development...
Metadata repository acts like a backbone to a data warehouse as it stores and manages the metadata that is the basis for all the operations of a data warehouse. The generalized metadata repository presented in this paper is a comprehensive approach for creating a data warehouse from multiple and heterogeneous data sources in a semi-automatic way. The approach addresses all the issues involved in fetching...
With the highly rapid increase of information, due to the development of massive application system, more and more people have an urgent need for a simple and rapid technology to integrate data which are stored in various data sources. The integration of heterogeneous data sources has become a central problem of modern computing. According to analyzing and researching general methodologies for heterogeneous...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.