Seminar 05/06 - Abstracts



The Computer Architect: A Unique Opportunity, A Unique Responsibility.
Yale Patt
University of Texas


Computer architecture is at the interface of what the software requires and what the hardware carries out. From one side, there are untapped opportunities for computers to solve problems. From the other side, the process technology people are promising 10 billion transistors on a die, operating at 10 GHz. That many transistors means even a single boring chip can have 200 Pentium 4 processors on it. We are in the middle -- what should we do. Some gurus have told us to do nothing, that all that can be done has been done. Others tell us we as a community need some shaking up. I think both are wrong, and that there is much we can do. I think there are areas that traditionally we have not thought of as within our sphere of influence where we could have important impact -- that is, each provides an opportunity, if we are willing to take on the accompanying responsibility. Algorithms, programming, education immediately come to mind. I might also talk about what I think future chcips might look like, if we have time.



Experimental Support for Reconfigurable Application-Specific Accelerators
Isaac Gelado
UPC-DAC


Abstract-- New computer architectures are being proposed and will be implanted in the next few years. A common trend to improve the system performance is to include some reconfigurable logic components into future multi-core chips. Several prototypes already exist but there still is a lack of support from the programming models, the compilers and the operating system. Reconfigurable architectures enable a new execution model based on hardware accelerators. There have been several efforts in the last few years to extend the thread abstraction to include the characteristics of these new reconfigurable elements. However, since the thread abstraction was conceived to abstract the CPU, the inherent model is different from what hardware accelerators should expect. In this paper we present a new way of managing hardware accelerators. We propose a new programming and execution model for hardware accelerators based on the processing unit abstraction. We study the relationship between hardware accelerators and the processor architecture, exploring the features that the new reconfigurable architectures should provide to the operating system. We also show how a prototype hardware accelerator achieves a 1.78x speed-up over the software implementation of the same functionality, out of a 41x potential speed-up, due to the data communication overhead. Finally we show our work on dynamic partial reconfiguration and how it enables using the processing unit abstraction to provide virtual reconfigurable logic.



Building a Global View for Optimization Purposes
Ramon Bertran
DAC-UPC


Since the approach of building specialized systems from existing general-purpose components is becoming more common, there is a growing need for global optimization tools that accomplish the functional requirements of these systems. The first approach for developers is to apply the current optimization techniques individually on each component. Nevertheless, these optimizations do not always improve the code enough and manual tuning must be done afterward. This paper presents a new approach for global optimization based on building a global view of the system, a global control flowgraph. The paper summarizes relevant characteristics of the system components to build a global view and presents particular examples for different architectures (PowerPC and x86) and operating systems (based on L4 kernel and Linux kernel). Our evaluations show that typical optimizations on a specific embedded environment reduce the code size by up to 54%.



Software Aging and Rejuvenation in SOAP-based Middleware Tools
Luis Silva
Univ. Coimbra, Portugal


Web-Services, SOAP and Service-oriented Architectures are gaining momentum in the area of distributed systems and internet applications. However, as we increase the abstraction level of the applications we are also increasing the complexity of the underlying middleware. In this talk, i will present a dependability benchmarking study to evaluate and compare the robustness of some of the most popular SOAP-RPC implementations that are intensively used in the industry. This benchmarking framework has support for different packages of middleware for client-server applications, like TCP/IP sockets, Java RMI, HTTP-XML and SOAP. The study was focused in detail on Apache Axis. The results of this benchmarking study should be seen as a contribution for the study of the impact of the complexity of the SOAP middleware in the dependability of SOA applications. In particular, we have observed a high susceptibility of software aging in this particular implementation of SOAP (Apache Axis1.3). Building on these results we propose a new SLA-oriented software rejuvenation technique that proved to be a simple way to increase the dependability of the SOAP-servers, the degree of self-healing and to maintain a sustained level of performance in the applications. Keywords : Dependability benchmarking; Dependable Web-Services; Self-healing systems; Software rejuvenation; Software aging; Autonomic Computing



Statistical Compiler Tuning
Masayo Haneda
Leiden Institute of Advanced Computer Science (LIACS)


Conventionally, programmers use standard -Ox settings which are provided by compiler developers. However, in order to obtain maximal performance, it is necessary to tune the compiler setting for the application as well as the underlying architecture. We propose a methodology to configure compiler options automatically using profile information. We apply statistical analysis to decide whether to turn on or to turn off compiler flags. This approach produces compiler settings of gcc3.3.1 for the SPEC2000 benchmark suite that outperform the O3 setting on three representative architectures: Pentium4, SPARC, and IA64.



Can Failure Prediction Methods Be Used for Performance Evaluation?
Miroslaw Malek
Institut für Informatik, Humboldt-Universitaet zu Berlin


The availability of software and hardware systems can be increased by preventive measures which are triggered by failure prediction mechanisms. We present and evaluate two techniques which model and predict the occurrence of failures as a function of discrete and continuous measurements of system variables. We employ two modelling approaches: an extended Markov chain model and a function approximation technique utilising universal basis functions (UBF). The presented modelling methods are data driven rather than analytical and can handle large amounts of variables and data. They offer the potential to capture the underlying dynamics of high-dimensional and noisy systems. Both modelling techniques have been applied to real data of a commercial telecommunication platform. The data includes event-based log files and measured system states. Results are presented in terms of precision, recall, F-Measure and cumulative cost. We compare our results to standard techniques such as linear ARMA models. Our findings suggest significantly improved forecasting performance compared to alternative approaches. By using the presented modelling techniques the system availability may be improved by an order of magnitude. We conjecture that the UBF approach can be used successfully for performance prediction in both single and multiprocessor systems.



Proposta de recerca en "Gestio Automatica de Consum Energetic" d'un computador des del Sistema Operatiu
Sergi Casas
Hewlett-Packard


Per diferents raons economiques i tambe ecologiques, minimitzar el consum energetic dels dispositius de computacio s'esta convertint en un avantatge competitiu a la industria, molt especialment a l'hora de vendre a grans comptes com governs i corporacions. El dispositiu ideal seria aquell que adaptes el seu consum energetic a la feina util que estigui realitzant en cada moment, tot evitant malbaratar energia. El Sistema Operatiu es troba en el millor punt per avaluar en cada moment les necessitats energetiques i per coordinar els canvis d'estat necessaris per a modificar dinamicament els perfils de consum, i l'electronica cada vegada ofereix mes punts de control per poder modular el consum. A la industria hi ha algunes aproximacions mes aviat barroeres, en parlarem i veurem les oportunitats per fer aportacions interessants en aquesta area



Assembly technology for parallel implementation of the large scale numerical models
Victor Malyshkin
Supercomputer Software Department, Institute of Computational Mathematics and Mathematical Geophysics (Computing Center) of the Siberian Division of the Russian Academy of Sciences (RAS)


One of the projects of SSD (Supercomputer Software Department) is presented. Assembly technology oriented to the development of high performance parallel programs of numerical modeling is discussed. The technology provides automatically such dynamic properties of application programs as dynamic tuning to the available resources, dynamic load balancing, etc. The technology application is demonstrated by the parallel implementation of the large scale model based on Particle-In-Cell (PIC) method. 1. About SSD and Chairs 2. Short review of the basic project of SSD 3. Methods of the PIC parallelization 4. Parallel PIC implementation 5. Modeling of Galaxy and Protoplanetary disc evolution



 

That's all folks!!!!!