Browsing by Author "Grossman, Max"
Now showing 1 - 3 of 3
Results Per Page
Sort Options
Item Productive Programming Systems for Heterogeneous Supercomputers(2017-01-31) Grossman, Max; Sarkar, VivekThe majority of today's scientific and data analytics workloads are still run on relatively energy inefficient, heavyweight, general-purpose processing cores, often referred to in the literature as latency-oriented architectures. The flexibility of these architectures and the programmer aids included (e.g. large and deep cache hierarchies, branch prediction logic, pre-fetch logic) makes them flexible enough to run a wide range of applications fast. However, we have started to see growth in the use of lightweight, simpler, energy-efficient, and functionally constrained cores. These architectures are commonly referred to as throughput-oriented. Within each shared memory node, the computational backbone of future throughput-oriented HPC machines will consist of large pools of lightweight cores. The first wave of throughput-oriented computing came in the mid 2000's with the use of GPUs for general-purpose and scientific computing. Today we are entering the second wave of throughput-oriented computing, with the introduction of NVIDIA Pascal GPUs, Intel Knights Landing Xeon Phi processors, the Epiphany Co-Processor, the Sunway MPP, and other throughput-oriented architectures that enable pre-exascale computing. However, while the majority of the FLOPS in designs for future HPC systems come from throughput-oriented architectures, they are still commonly paired with latency-oriented cores which handle management functions and lightweight/un-parallelizable computational kernels. Hence, most future HPC machines will be heterogeneous in their processing cores. However, the heterogeneity of future machines will not be limited to the processing elements. Indeed, heterogeneity will also exist in the storage, networking, memory, and software stacks of future supercomputers. As a result, it will be necessary to combine many different programming models and libraries in a single application. How to do so in a programmable and well-performing manner is an open research question. This thesis addresses this question using two approaches. First, we explore using managed runtimes on HPC platforms. As a result of their high-level programming models, these managed runtimes have a long history of supporting data analytics workloads on commodity hardware, but often come with overheads which make them less common in the HPC domain. Managed runtimes are also not supported natively on throughput-oriented architectures. Second, we explore the use of a modular programming model and work-stealing runtime to compose the programming and scheduling of multiple third-party HPC libraries. This approach leverages existing investment in HPC libraries, unifies the scheduling of work on a platform, and is designed to quickly support new programming model and runtime extensions. In support of these two approaches, this thesis also makes novel contributions in tooling for future supercomputers. We demonstrate the value of checkpoints as a software development tool on current and future HPC machines, and present novel techniques in performance prediction across heterogeneous cores.Item Programming Models and Runtimes for Heterogeneous Systems(2013-09-16) Grossman, Max; Sarkar, Vivek; Mellor-Crummey, John; Cox, Alan L.With the plateauing of processor frequencies and increase in energy consumption in computing, application developers are seeking new sources of performance acceleration. Heterogeneous platforms with multiple processor architectures offer one possible avenue to address these challenges. However, modern heterogeneous programming models tend to be either so low-level as to severely hinder programmer productivity, or so high-level as to limit optimization opportunities. The novel systems presented in this thesis strike a better balance between abstraction and transparency, enabling programmers to be productive and produce high-performance applications on heterogeneous platforms. This thesis starts by summarizing the strengths, weaknesses, and features of existing heterogeneous programming models. It then introduces and evaluates four novel heterogeneous programming models and runtime systems: JCUDA, CnC-CUDA, DyGR, and HadoopCL. We'll conclude by positioning the key contributions of each piece in this thesis relative to the state-of-the-art, and outline possible directions for future work.Item The Texas flood registry: a flexible tool for environmental and public health practitioners and researchers(Springer Nature, 2021) Miranda, Marie Lynn; Callender, Rashida; Canales, Joally M.; Craft, Elena; Ensor, Katherine B.; Grossman, Max; Hopkins, Loren; Johnston, Jocelyn; Shah, Umair; Tootoo, JoshuaBackground: Making landfall in Rockport, Texas in August 2017, Hurricane Harvey resulted in unprecedented flooding, displacing tens of thousands of people, and creating environmental hazards and exposures for many more. Objective: We describe a collaborative project to establish the Texas Flood Registry to track the health and housing impacts of major flooding events. Methods: Those who enroll in the registry answer retrospective questions regarding the impact of storms on their health and housing status. We recruit both those who did and did not flood during storm events to enable key comparisons. We leverage partnerships with multiple local health departments, community groups, and media outlets to recruit broadly. We performed a preliminary analysis using multivariable logistic regression and a binomial Bayesian conditional autoregressive (CAR) spatial model. Results: We find that those whose homes flooded, or who came into direct skin contact with flood water, are more likely to experience a series of self-reported health effects. Median household income is inversely related to adverse health effects, and spatial analysis provides important insights within the modeling approach. Significance: Global climate change is likely to increase the number and intensity of rainfall events, resulting in additional health burdens. Population-level data on the health and housing impacts of major flooding events is imperative in preparing for our planet’s future.