The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
Parallel programming has been an active area of research in computer science and software engineering for many years. Parallel programming should ideally provide a linear speedup to computational problems. In reality, this is rarely the case. While there are some algorithms that cannot be parallelized, many that can, still fail to provide the ideal linear speedup. For algorithms that can benefit from...
In this paper, we present our Concurrent Systems class, where parallel programming and parallel and distributed computing (PDC) concepts have been taught for more than 20 years. Despite several rounds of changes in hardware, the class maintains its goals of allowing students to learn parallel computer organizations, studying parallel algorithms, and writing code to be able to run on parallel and distributed...
In this paper, the performance of parallel computing will be thoroughly discussed in the domain of image matching. The concept of image matching is widely used in the areas of security, medical and computer vision which require comparing two images for similarities. However, depending on the size of images, it is highly possible that the application computation cannot be handled in a single processor...
This work describes research, efforts, and outcomes for several Computer Science courses after incorporating XSEDE High Performance Computing (HPC) Resources and recommended and required curriculum additions from the ACM 2013 Computer Science Curricula and IEEE Technical Committee on Parallel Processing Curriculum Guidelines. The work herein describes the courses affected by this work, including Computer...
As a deterministic parallel programming language that guarantees safety of parallel executions, EFL was designed to allow the embedding of parallel code blocks into a sequential host language program. An EFL pre-compiler, which is described here, has been implemented that translates EFL blocks into the host language. The EFL pre-compiler and runtime supports parallel building blocks such as parallel...
Convolution computing plays an important role in scientific computing. However, traditional Message Passing Interface (MPI) model has the disadvantages such as massive message passing and load-unbalancing. According to these problems, this paper propose a new parallel convolution based on MPI model, which is able to effectively balance the load as well as bring great reduction on message passing....
Most people, at one point or another, have wondered at the extent of the capabilities of the materials they currently use. To some, this means establishing a point where they venture no farther -- a limit, if you will. To others, this is the start of where find out just where they can push the boundaries. Almost every multi-core, multiprocessor or multiple unit computer setups these days largely lack...
Along with the comprehensive uses of parallel computings in various fields at present, clusters have been proved to be the main trend of the parallel computing systems beyond question. Compared with the traditional specialized clusters, the universal clusters constructed with PC costed down markedly. This paper focused on how to build a diminutive isomorphic PC cluster based on the Linux operating...
Efficient parallel programming has always been very tricky and only expert programmers are able to take the most of the computing power of modern computers. Such a situation is an obstacle to the development of the high performance computing in other sciences as well as in the industry. The fast changes in the computer architecture (multicores, manycores, GPU, clusters, …) make even more difficult,...
The abundant availability of multi-core computers makes "parallel computers" a common place and teaching Computer Science students to be able to design and develop parallel algorithms an urgent task. Most students recognize the needs of developing skills in parallel programming. However, since their Computer Science related curriculum are mostly taught based on sequential computers, introducing...
Today, almost all desktop and laptop computers are shared-memory multicores, but the code they run is overwhelmingly serial. High level language extensions and libraries (e.g., Open MP, Cilk++, TBB) make it much easier for programmers to write parallel code than previous approaches (e.g., MPI), in large part thanks to the efficient {\em work-stealing} scheduler that allows the programmer to expose...
Genetic algorithms is facing the low evolution rate and difficulties to meet real-time requirements when handing large-scale combinatorial optimization problems. In this paper, we propose a coarse-grained-master-slave hybrid parallel genetic algorithm model based on multi-core cluster systems. This model integrates the message-passing model and the shared-memory model. We use message-passing model—MPI...
This paper proposes parallel algorithm of a general three-phase load flow calculation based on symmetrical components. The proposed algorithm comprises of two part i.e. sequence admittance matrix construction and sequence decoupled algorithm for sequence networks solver which have three independent sub-problems corresponding to positive, negative, and zero-sequence networks. These two parts are solved...
The idea of distributed computer emulation is presented within this paper. Since classic emulation techniques put the power load on the host CPU only, the new approach tries to distribute the load among other available processors within the host platform. The implementation uses OpenCL framework. This standard allows writing high parallel and portable programs in ISO C99 subset language, runnable...
Introduces the design and implementation of a Remote System for Parallel Computing.Based on Multi-platform.This system support remote operating and running Linux system parallel programes on Windows system and provide a kind of uniform mode for the applied customer to operation, The customer does not need to know the concrete structure of the parallel computing system,can operate the parallel system...
The long computation time caused by sequential processing of huge size image in three-dimensional image direct writing prevents its industrialization. To reduce computation time, OpenMP technique is applied to process the images in a parallel way. Performance improves by both parallel image reading and parallel image processing. Experiment results show that the method possesses high effectiveness...
Summary form only given. Chip multiprocessors (CMPs) are now the dominant architecture in microprocessor design. However, in many software environments, due to the difficulty of writing correct and high performing parallel programs, the capability of CMPs is underutilized. In this talk, I will argue that to enable more parallel programs to be written more easily, the design of CMPs should be driven...
With the development of parallel technology and application, the performance analysis and visualization of parallel computing is one of the most important parts. This paper presents a parallel computing performance visualization system which is developed to visualize and analyze the speedup and efficiency of parallel computing. We visualize and analyze the results of magnetotelluric parallel forward...
Efficient mapping of logical processes to physical processes is one of key technologies to accelerate parallel performance simulation. Aiming at minimizing the communications between SMP nodes and between host physical processes, this paper presents a novel method named TPsmp-LP3M. It automatically extracts communication pattern of logical processes from trace and then generates a two-phase mapping...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.