The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
In this paper a log-MAP turbo decoding algorithm which incorporates reliability threshold based trellis branch elimination together with mean based early iteration termination has been proposed. The proposed algorithm reduces computational complexity by eliminating branches in trellis. Since the minimum Log Likelihood Ratio (LLR) obtained for each iteration varies with channel conditions, an early...
A novel approach of a testbed for embedded networking nodes has been conceptualized and implemented. It is based on the use of virtual nodes in a PC environment, where each node executes the original embedded code. Different nodes are running in parallel and are connected via so-called virtual interfaces. The presented approach is very efficient and allows a simple description of test cases without...
Semantic web services represent the potential of the web and they have significant impact on the discovery process. Due to the high proliferation of web services, selecting the best web services from functional equivalent service providers have become a real challenge when a large number of services have been published in a registry. If these services have been functionally-equivalent, it is difficult...
This paper presents an effective design space exploration strategy for the development of dependable systems using selective hardening techniques based on software. Instead of design space exploration approaches based on brute-force or time-consuming fault injection experiments, this strategy is grounded in an early estimation of the register file criticality in microprocessor-based systems. This...
Accurate measurement of the quality of systems is crucial to building trustworthy systems. Such a measurement indicates whether a system is working properly and meeting its requirements. Although security and dependability metrics are regarded as key metrics for measuring the quality of systems, they are not sufficient for measuring the quality of systems that are placed in a multi-domain environment...
The process of testing any software system is an atrocious task which indeed consumes a ton of effort, and expensive also. Required effort and time to do adequate as well as effective testing get bigger, as the software gets more complexed that can lead to swarm over the project budget or some test cases left uncovered or delay in completion. A suitably generated test suite does not only locate errors...
Software development is an essential field today. The advancement in software systems leads to risk of them being exposed to defects. It is important to predict the defects well in advance in order to help the researchers and developers to build cost effective and reliable software. Defect prediction models extract information about the software from its past releases and predict the occurrence of...
Design by Contract (DbC) is a software engineering paradigm that aims to reduce the amount of source code in a computer program by avoiding redundant parameter checking. It is based on strict interface specifications, including the specification of responsibilities for checking. Although DbC potentially leads to better quality programs, it has to be supported by the programming language/environment...
Research on similarity searching of cheminformatic data sets has been focused on similarity measures using fingerprints. However, nominal scales are the least informative of all metric scales, increasing the tied similarity scores, and decreasing the effectivity of the retrieval engines. Tanimoto's coefficient has been claimed to be the most prominent measure for this task. Nevertheless, this field...
In this paper we take a closer look at the operation of software defined networking (SDN) in intra-domain networks. The focus is on the dependability issues related to interworking of SDN controllers, network OS (NOS), and forwarding in the data plane. Both the separation of the control and data planes, and the (virtually) centralized control processes, are challenging from a dependability perspective...
Debugging is crucial for producing reliable software. One of the effective bug localization techniques is Spectral-Based Fault Localization (SBFL). It locates a buggy statement by applying an evaluation metric to program spectra and ranking program components on the basis of the score it computes. Recently, genetic programming has been proposed as a way to find good metrics. We have found that the...
Agile Software Development (ASD) has been on mainstream through methodologies such as XP and Scrum enabling them to be applied in the development of complex and reliable software systems. This paper is the end result of the Master's dissertation of the main author, and proposes a solution to guide the development of complex systems based on components by adding exceptional behavior modeling practices...
Architectural qualities such as reliability, performance, and security, are often realized in a software system through the adoption of tactical design decisions such as the decision to use redundant processes, a heartbeat monitor, or a specific authentication mechanism. Such decisions are critical for delivering a system that meets its quality requirements. Despite the stability of high-level decisions,...
This paper analyses the nature of fault tolerance software-based techniques and the influence of their overheads to determine an efficient strategy for applying those techniques in a selective way. Several considerations that have to be taken into account are presented in this work. These include an analysis of fault coverage and overheads when selective hardening is adopted; side effects of selective...
This paper presents a method for early software reliability assessment with planning and measurement of achieved reliability improvement during the development process of the software. Early assessment models for software (SW) reliability improvement contain parameters which are estimated using the existing records of reliability improvements, as related to the software size and applied SW development...
Using design models instead of executable code has shown itself to be an efficient way of increasing abstraction level of software development. However, applying established code-based software engineering methods to design models can be a challenge - due to different abstraction levels, the same metrics as for code are not applicable for the design models. One of practical challenges in using metrics...
Exception handling mechanisms (EHM) were conceived as a means to improve maintainability and reliability of programs that have to deal with exceptional situations. Amongst the different implementations of built-in EHM, we classify them in two main categories: reliability-driven and maintenance-driven. Some programming languages, such as Java, provide built-in exception handling mechanisms that promote...
The functionality and the performance of smart environment applications can be hampered by faults. Fault tolerance solutions aim to achieve graceful performance degradation in the presence of faults, ideally without leading to application failures. This is a reactive approach and, by itself, gives little flexibility and time for preventing potential failures. We propose a proactive fault-prevention...
The radio frequency (RF) spectrum is a limited resource. Spectrum allotment disputes stem from this scarcity as many radio devices are confined to a fixed frequency or frequency sequence. One alternative is to incorporate cognition within a reconfigurable radio platform, therefore enabling the radio to adapt to dynamic RF spectrum environments. In this way, the radio is able to actively sense the...
A cyber-physical system (CPS) is a system featuring a tight combination of, and coordination between, the system's computational and physical elements. System reliability is a critical requirement of cyber-physical systems. An unreliable CPS often leads to system malfunctions, service disruptions, financial losses and even human life. Improving CPS reliability requires an objective measurement, estimation...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.