Languages and Compilers for Parallel Computing

Languages and Compilers for Parallel Computing
-0 %
22nd International Workshop, LCPC 2009, Newark, DE, USA, October 8-10, 2009, Revised Selected Papers
Besorgungstitel - wird vorgemerkt | Lieferzeit: Besorgungstitel - Lieferbar innerhalb von 10 Werktagen I

Unser bisheriger Preis:ORGPRICE: 96,29 €

Jetzt 53,48 €*

Alle Preise inkl. MwSt. | Versandkostenfrei
Artikel-Nr:
9783642133732
Veröffentl:
2010
Seiten:
426
Autor:
Guang R. Gao
Gewicht:
661 g
Format:
237x190x20 mm
Serie:
5898, Theoretical Computer Science and General Issues
Sprache:
Englisch
Beschreibung:

Itisourpleasuretopresentthepapersacceptedforthe22ndInternationalWo- shop on Languages and Compilers for Parallel Computing held during October 8-10 2009 in Newark Delaware, USA. Since 1986, LCPC has became a valuable venueforresearchersto reportonworkinthegeneralareaofparallelcomputing, high-performance computer architecture and compilers. LCPC 2009 continued this tradition and in particular extended the area of interest to new parallel computing accelerators such as the IBM Cell Processor and Graphic Processing Unit (GPU). This year we received 52 submissions from 15 countries. Each submission receivedatleastthreereviewsandmosthadfour.ThePCalsosoughtadditional externalreviewsforcontentiouspapers.ThePCheldanall-dayphoneconference on August 24 to discuss the papers. PC members who had a con?ict of interest were asked to leave the call temporarily when the corresponding papers were discussed. From the 52 submissions, the PC selected 25 full papers and 5 short paperstobeincludedintheworkshopproceeding,representinga58%acceptance rate. We were fortunate to have three keynote speeches, a panel discussion and a tutorial in this year's workshop. First, Thomas Sterling, Professor of Computer Science at Louisiana State University, gave a keynote talk titled "HPC in Phase Change: Towards a New Parallel Execution Model." Sterling argued that a new multi-dimensional research thrust was required to realize the design goals with regard to power, complexity, clock rate and reliability in the new parallel c- puter systems.ParalleX,anexploratoryexecutionmodeldevelopedbySterling's group was introduced to guide the co-design of new architectures, programming methods and system software.
A Communication Framework for Fault-Tolerant Parallel Execution.- The STAPL pList.- Hardware Support for OpenMP Collective Operations.- Loop Transformation Recipes for Code Generation and Auto-Tuning.- MIMD Interpretation on a GPU.- TL-DAE: Thread-Level Decoupled Access/Execution for OpenMP on the Cyclops-64 Many-Core Processor.- Mapping Streaming Languages to General Purpose Processors through Vectorization.- A Balanced Approach to Application Performance Tuning.- Automatically Tuning Parallel and Parallelized Programs.- DFT Performance Prediction in FFTW.- Safe and Familiar Multi-core Programming by Means of a Hybrid Functional and Imperative Language.- Hierarchical Place Trees: A Portable Abstraction for Task Parallelism and Data Movement.- OSCAR API for Real-Time Low-Power Multicores and Its Performance on Multicores and SMP Servers.- Programming with Intervals.- Adaptive and Speculative Memory Consistency Support for Multi-core Architectures with On-Chip Local Memories.- Synchronization-Free Automatic Parallelization: Beyond Affine Iteration-Space Slicing.- Automatic Data Distribution for Improving Data Locality on the Cell BE Architecture.- Automatic Restructuring of Linked Data Structures.- Using the Meeting Graph Framework to Minimise Kernel Loop Unrolling for Scheduled Loops.- Efficient Tiled Loop Generation: D-Tiling.- Effective Source-to-Source Outlining to Support Whole Program Empirical Optimization.- Speculative Optimizations for Parallel Programs on Multicores.- Fastpath Speculative Parallelization.- PSnAP: Accurate Synthetic Address Streams through Memory Profiles.- Enforcing Textual Alignment of Collectives Using Dynamic Checks.- A Code Generation Approach for Auto-Vectorization in the Spade Compiler.- Portable Just-in-Time Specialization of DynamicallyTyped Scripting Languages.- Reducing Training Time in a One-Shot Machine Learning-Based Compiler.- Optimizing Local Memory Allocation and Assignment through a Decoupled Approach.- Unrolling Loops Containing Task Parallelism.

Kunden Rezensionen

Zu diesem Artikel ist noch keine Rezension vorhanden.
Helfen sie anderen Besuchern und verfassen Sie selbst eine Rezension.