The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
With the rapid development of database technique, categorizing datasets becomes very important for discovering information. Decision tree classification provides a rapid and effective method of categorizing datasets. Although many algorithmic methods exist for optimizing decision tree structure, these can be vulnerable to changes in the training dataset. In this paper, an evolutionary method is presented,...
There have been proposed spoken dialog systems that utilizes simple database consisted of example sentences and the corresponding reply sentences. However, it is costly to prepare this database manually. In the present study, we propose a framework in which both the example and reply sentences are automatically generated from a database description table that describes minimum information for describing...
Wi-Fi fingerprinting is a technique which can provide location in GPS-denied environments, relying exclusively on Wi-Fi signals. It first requires the construction of a database of “fingerprints”, i.e. signal strengths from different access points (APs) at different reference points in the desired coverage area. The location of the device is then obtained by measuring the signal strengths at its location,...
Tabu Search is a meta-heuristic approach successfully used to address optimization problems in several contexts. This paper reports the results of an empirical study carried out to investigate the effectiveness of Tabu Search in estimating Web application development effort. The dataset employed in this investigation is part of the Tukutuku database. This database has been used in several studies...
Computational approaches have been applied in many different biology application domains. When such tools are based on conventional computation, they have shown limitations to approach complex biological problems. In the present study, a computational evolutionary environment (CEE) is proposed as tool to extract classification rules from biological datasets. The main goal of the proposed approach...
Condition-based maintenance (CBM) is a maintenance approach wherein equipment repair or replacement decisions are based on the current and projected health of the equipment measured by periodic collection and analysis of data. In this context, the accuracy of data is vital. Unfortunately, missing and inaccurate data are recurring problems in many CBM database. These problems can cause bias or lead...
In order to promote the automation and intellectualization of e-government archiving, this article obtains from the e-government archiving research background to propose the DURA model, carries on the discussion to its integrant parts and the reciprocity. And it analyzes an essential factor - - the origin, the connotation and the integrated development environment of business rules (R) and the work...
Discrimination of machine-printed and hand-written words is deemed as a major problem in the recognition of the mixed texts. To present a new method to distinguish between machine-printed words and hand-written words using a novel statistical feature on base legibility and discriminator threshold are objectives of this study. Because of the hand trembling, sudden uncontrollable movement of hand and...
With the problem that miss and false retrieval in the technical article retrieval platform, this paper brings a new retrieval method for the forestry technical articles. The specialized vocabulary database is constructed to store all the key words for Forestry Journals in Chinese. The retrieval method firstly retrieves the key words from the different word stocks, and gets all the synonyms form them,...
Inferring PoP level maps is gaining interest due to its importance to many areas, e.g., for tracking the Internet evolution and studying its properties. In this paper we introduce a novel structural approach to automatically generate large scale PoP level maps using traceroute measurement from multiple locations. The PoPs are first identified based on their structure, and then are assigned a location...
Geographic location and Grid computing are two areas that have taken off in recent years, both receiving a lot of attention from research community. The Grid Resource Brokers, which tries to find the best match between the job requirements and the resources available on the Grid, can take benefits by knowing the geographic location of clients, for a considerable improvement of their decision-taking...
Recently, some accidents which are caused by external corrosion of pipes often occur in chemical and petrochemical plants. It is very difficult to detect points of the pipes where corrosion has considerably developed because the pipes are usually covered with insulations and cannot be inspected directly. Moreover, the length of the pipes is more than 1000 km, then it needs huge time and cost to inspect...
Result Validation is an important phase of guidance simulation VV&A process, which main purpose is testing the consistency of simulation data and flight data or other standard data. Result Validation mainly involves the data analysis and calculate activities, therefore, developing the result validation tool (RVT), make use of computer's computing instead of manual calculation, can consumedly increase...
In order to improve the efficiency of database management and intelligence of database record classification, a classification method of network database record based on fuzzy theory is proposed in this paper. Firstly, an automatic classification frame of database is constructed, and then standard record model and special data record and new record model on fuzzy set are given. By calculating the...
In unit selection based text-to-speech (TTS) synthesis, the accurate position of the unit boundaries in the unit selection database is one of the factors that determine the quality of the synthesized speech. To ensure the accuracy of the boundary positions, developers often have to manually verify the speech boundaries that are generated by automatic speech recognition techniques. In order to reduce...
Biometric identification has lately attracted attention because of its high convenience; it does not require a user to enter a user ID. The identification accuracy, however, degrades as the number of the enrollees increases. Although many multimodal biometric techniques have been proposed to improve the identification accuracy, it requires the user to input multiple biometric samples and makes the...
Many organizations collect large amounts of data to support their business and decision making processes. The data collected from various sources may have data quality problems in it. These kinds of issues become prominent when various databases are integrated. The integrated databases inherit the data quality problems that were present in the source database. The data in the integrated systems need...
An improved K-medoids clustering algorithm (IKMC) to resolve the problem of detecting the near-duplicated records is proposed in this paper. It considers every record in database as one separate data object, uses edit-distance method and the weights of attributes to get similarity value among records, then detect duplicated records by clustering these similarity value. This algorithm can automatically...
In this paper we present a report on the data quality of a medical database containing clinical and administrative data from hospitals and private clinics in the Bologna district area. In particular, we have analyzed the database according to several data quality dimensions, identifying a number of issues (e.g., inaccuracies and incompleteness) that we have systematized and described in this work.
The information about which modules in a software system's future version are potentially defective is a valuable aid for quality managers and testers. Defect prediction promises to indicate these defect-prone modules. Constructing effective defect prediction models in an industrial setting involves the decision from what data source the defect predictors should be derived. In this paper we compare...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.