With an increasing amount of data and more complex algorithms available to scientists and practitioners today, parallel processing is almost always a must, and in fact, is expected in packages implementing time-consuming methods. This course introduces you to concepts and tools available in R for parallel computing and provides solutions to a few important non-trivial issues in parallel processing like reproducibility, generating random numbers and load balancing.
Can I Run My Application in Parallel?Free
In order to take advantage of parallel environment, the application needs to be split into pieces. In this introductory chapter, you will learn about different ways of partitioning and how it fits different hardware configurations. You will also be introduced to various R packages that support parallel programming.
This chapter will dive deeper into the parallel package. You'll learn about the various backends and their differences and get a deep understanding about the workhorse of the package, namely the clusterApply() function. Strategies for task segmentation including their pitfalls will also be discussed.
foreach, future.apply and Load Balancing
In this chapter, you will look at two user-contributed packages, namely foreach and future.apply, which make parallel programming in R even easier. They are built on top of the parallel and future packages. In the last lesson of this chapter, you will learn about the advantages and pitfalls of load balancing and scheduling.
Random Numbers and Reproducibility
Now you might ask, can I reproduce my results if the application uses random numbers? Can I generate the same results regardless of if the code runs sequentially or in parallel? This chapter will answer these questions. You will learn about a random number generator well suited to a parallel environment and how the various packages make use of it.
Senior Research Scientist, University of Washington
Hana works as a senior research scientist at the University of Washington’s Center for Statistics and the Social Sciences, in the area of statistical computing. She also works as a data scientist at the Puget Sound Regional Council in Seattle. She has been involved in projects advancing parallel computing in statistics. She contributed to the snow package which became the R core package parallel. She developed a methodology for fault tolerant and reproducible parallel computing, implemented in snowFT, as well as the first interface to the L’Ecuyer's random number generator, the rlecuyer package. Currently, she’s been working in the area of statistical demography, developing R packages used by the United Nations for projecting demographic indicators for all countries of the world.