The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
In many areas of computer architecture design and program development, the knowledge of dynamic program behavior can be very handy. Several challenges beset the accurate and complete collection of dynamic control flow and memory reference information. These include scalability issues, runtime-overhead, and code coverage. For example, while Tallam and Gupta's work on extending WPP (Whole Program Paths)...
We develop a new metric for job scheduling that includes the effects of memory contention amongst simultaneously-executing jobs that share a given level of memory. Rather than assuming each job or process has a fixed, static memory requirement, we consider a generalscenario wherein a process’ performance monotonically increases as a function of allocated memory, as defined by a miss-rate versus memory...
The goal of the Ultra-Scale Computing Valuation Project is to understand utilization issues for both users and managers of the largest scienti.c computing systems and to begin developing appropriate metrics and models for such system. This paper describes a few aspects of the project.
Theoretical research on parallel algorithms has focused on NC theory. This motivates the development of parallel algorithms that are extremely fast, but possibly wasteful in their use of processors. Such algorithms seem of limited interest for real applications currently run on parallel computers. This paper explores an alternative approach that emphasizes the efficiency of parallel algorithms. We...
The popularity of research on the scheduling of parallel jobs demands a periodic review of the status of the field. Indeed, several surveys have been written on this topic in the context of parallel supercomputers [17, 20]. The purpose of the present paper is to update that material, and to extend it to include work concerning clusters and the grid.
We describe the design of a communications board that supports collective communications operations of parallel programs. Each cluster of processors are interconnected by a bus and connected to the communications board. The communications boards are, in turn, connected to a low latency, high bandwidth, slow configuration time crossbar network. The design is geared towards low-end (inexpensive) parallel...
Consider a scenario in which a smart phone automatically saves the user’s positional records for personalized location-based applications. The smart phone will infer patterns of user activities from the historical records and predict user’s future movements. In this paper, we present algorithms for mining the evolving positional logs in order to identify places of significance to user and representative...
The job workloads of general-purpose multiprocessors usually include both compute-bound parallel jobs, which often require gang scheduling, as well as I/O-bound jobs, which require high CPU priority for the individual gang members of the job in order to achieve interactive response times. Our results indicate that an effective interactive multiprocessor scheduler must be flexible and tailor the priority,...
Parallel job scheduling is beginning to gain recognition as an important topic that is distinct from the scheduling of tasks within a parallel job by the programmer or runtime system. The main issue is how to share the resources of the parallel machine among a number of competing jobs, giving each the required level of service. This level of scheduling is done by the operating system. The four most...
This paper presents a new weight incidence representation of Dynamic wavelength addressing in optical fiber networks utilizing wavelength division multiplexing (WDM) can form the basis for a high-performance, high-bandwidth, low-latency any-to-any interconnection network. WDM optical fiber networks exploit the fact that photons of different wavelengths do not normally interact, thereby enabling the...
Application debugging is a tedious but inevitable chore in any software development project. An effective debugger can make programmers more productive by allowing them to pause execution and inspect the state of the process, or monitor writes to memory to detect data corruption. This paper introduces the new concept of Efficient Debugging using Dynamic Instrumentation (EDDI). The paper demonstrates...
The space of job schedulers for parallel supercomputers is rather fragmented, because different researchers tend to make different assumptions about the goals of the scheduler, the information that is available about the workload, and the operations that the scheduler may perform. We argue that by identifying these assumptions explicitly, it is possible to reach a level of convergence. For example,...
The evaluation of parallel job schedulers hinges on two things: the use of appropriate metrics, and the use of appropriate workloads on which the scheduler can operate. We argue that the focus should be on on-line open systems, and propose that a standard workload should be used as a benchmark for schedulers. This benchmark will specify distributions of parallelism and runtime, as found by analyzing...
For the past six months, I have been integrating several experimental, cuttingedge technologies developed by my colleagues at MIT as part of the MIT LCS/ AIL Oxygen project. This paper gives a snapshot of this work-in-progress.
Despite the large number of papers that have been published, scheduling and load balancing continue to be an active area of research.The topic covers all aspects related to scheduling and load balancing including application and system level techniques, theoretical foundations and practical tools.New aspects of parallel and distributed systems, such as clusters, grids, and global computing require...
Gang scheduling is a resource management scheme for parallel and distributed systems that combines time-sharing and space-sharing to ensure high overall system throughput and short response times for interactive tasks. We recently participated in the design and implementation of a flexible gang scheduling scheme on an IBM SP2 parallel system and a cluster of IBM RS/6000 workstations. In this paper,...
Envelopes, a run-time mechanism which automatically supports adaptive local queues for MIMD load balancing, are proposed and demonstrated. Envelopes promote generality and language simplicity, while sustaining efficiency. The local queues, one for each PE, contain a get_work task which pulls activities from a global list. In addition, they contain one or more envelopes within which activities...
The scheduling of jobs on parallel supercomputer is becoming the subject of much research. However, there is concern about the divergence of theory and practice. We review theoretical research in this area, and recommendations based on recent results. This is contrasted with a proposal for standard interfaces among the components of a scheduling system, that has grown from requirements in the field.
This book constitutes the strictly refereed post-workshop proceedings of the 1997 IPPS Workshop on Job Scheduling Strategies for Parallel Processing held in Geneva, Switzerland, in April 1997, as a satelite meeting of the IEEE/CS International Parallel Processing Symposium.The 12 revised full papers presented were carefully reviewed and revised for inclusion in the book. Also included is a detailed...
This volume contains the papers selected after a very careful refereeing process for presentation during the Workshop on Job Scheduling Stategies for Parallel Processing, held in Santa Barbara, California, as a prelude to the IPPS '95 conference in April 1995.The 19 full papers presented demonstrate that parallel job scheduling takes on a crucial role as multi-user parallel supercomputers become more...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.