Notes:

1. The entire source code of this post in here

2. The PDF version of this post in here

In previous two blogs (here and here), we illustrated several skills to build and optimize artificial neural network (ANN) with R and speed up by parallel BLAS libraries in modern hardware platform including Intel Xeon and NVIDIA GPU. Nowadays, multiple GPU accelerations are crucial for learning huge networks, one example, as Microsoft won ImageNet competition with huge network up to 1000 layers in 2015, [here and here]. In this blog, I will focus on applying CUDA implementation into our neural network offloading the computationally intensive parts into GPU and then we can easily extend CUDA implementation from single GPU to multiple GPUs under ‘parallel’ packages of R. P.S: If you want to go through all of these contents quickly, check out my presentation in GTC16 in here.

## CUDA INTEGRATION

Now, we begin to introduce our ninja skills : CUDA

After combined DNN code with CUDA BLAS library and several optimizations, we get the follow results in the table in the previous blog and leave one question for readers:

**What is your opinion about the next step of optimizations?**

It’s obvious that the function ‘pmax’ accounts for lots of runtimes (**31.58** secs) following ‘%*%’ (**53.72** secs) since ‘pmax’ is implemented by R and it will be very slow when the data size increase. (btw, you can try to increase the number of neurons in hidden layer to 1024 and profiling the code again to figure out the ratio of ‘pmax’ in the whole computation). Reviewing the functionality of ‘pmax’ in our DNN case, we implement the ReLU function and get the maximum value among input value and 0 for every neuron in hidden layer. Furthermore, because our algorithm is vectorized by matrices for high performance, the input of ‘pmax’ is a two-dimensional matrix and we can parallel maximum function into each element easily. So, let’s start to parallel the ReLu function by CUDA. I will skip the details of CUDA programming in this blog, you can refer programming guide in here.

First, we can write the CUDA code to compare the input value with ZERO. Despite it is a very naïve implementation, it is still very fast than original R code. Throughout this kernel code can be executed in NVIDIA GPU, it is written by CUDA C rather than R. Next, we need to call it from R environment.

In general, we have to write wrapper functions to bridge R, C/C++, and CUDA. Prof. Matloff and I have written the blog introduced linking R with CUDA step by step regarding with .Call() and .C() function with two-level wrappers from R to C/C++ and C/C++ to CUDA (here and here). A brief summary about the major difference of .C() and .Call() is shown in below table.

From the performance view, the .Call() function is selected since little overhead between R and C/C++ by avoiding explicit copying data from R to C/C++ and then from C/C++ back to R. In below code, we access data by a pointer and heavily use R internal structures with very efficient way.

// difinition for R extern "C" { SEXP pmax_cuda(SEXP A, SEXP threshold, SEXP devID); } //CUDA: simple implementation of pmax __global__ void pmax_kernel(double *A, const int M, const int N, const double threshold) { int tid = blockIdx.x * blockDim.x + threadIdx.x; if(tid < M*N) { A[tid] = (A[tid]>threshold)?A[tid]:0; } return; } // Wrapper code between R and CUDA SEXP pmax_cuda(SEXP A, SEXP threshold, SEXP devID) { // data structure for GPU double *A_host = NULL; double *A_d = NULL; double gw = 0; int mm = 0, nn = 0; int gpuID = 0; // data transfer from R to C by pointers A_host = REAL(A); SEXP Rdim = getAttrib(A, R_DimSymbol); mm = INTEGER(Rdim)[0]; nn = INTEGER(Rdim)[1]; gw = REAL(threshold)[0]; gpuID = INTEGER(devID)[0]; // for multiple GPU case cudaSetDevice(gpuID); // return value, allocated in C and can be used in R directly SEXP Rval; PROTECT(Rval = allocVector(REALSXP, mm*nn)); // GPU memory allocation cudaMalloc(&A_d, mm*nn*sizeof(double)); if(NULL == A_d) { printf("\nNo RAM space in GPU!\n"); UNPROTECT(1); return R_NilValue; } // memory copy from CPU to GPU cudaMemcpy(A_d, A_host, mm*nn*sizeof(double), cudaMemcpyHostToDevice); // CUDA: pmax, really computation parts pmax_kernel<<<(mm*nn-1)/512+1, 512>>>(A_d, mm, nn, gw); cudaMemcpy(REAL(Rval), A_d, mm*nn*sizeof(double), cudaMemcpyDeviceToHost); cudaDeviceSynchronize(); // Free unused memory of GPU if(A_d) {cudaFree(A_d); A_d=NULL;} UNPROTECT(1); return Rval; }

Next, compile the C/C++ and CUDA code together to a shared object file (.so) or dynamic link library (.dll) for loading in R.

nvcc -O3 -arch=sm_35 -G -I…/CUDA-v7.5.18/include -I…/R-3.2.0/bin/lib64/R/include/ -L…/R/lib64/R/lib –shared -Xcompiler -fPIC -o cudaR.so cudaR.cu

Finally, the CUDA version of ‘pmax’ can be called in R as simple as R builtin function with R’s wrapper, and, for infrastructure engineer, writing a nice wrapper is still an important job 🙂

Show our performance now! By replacing ‘pmax’ with new ‘pmax.cuda’, the execution time of pmax reduces to **6.7** seconds from original 31.58 so it’s **5X speedup** and totally the** 1.2X speedup** gains.

**Scale out to MultiGPU**

Parallel computing is not a novel concept in R community. Data scientist is familiar with parallel strategies both in speedup their model construction and inference. In fact, the requirement of parallel computing in R is even higher than C/C++. The C/C++ implementations always focus on low-level instructions and optimizations such as memory locality, communication, computation efficiency and much more while R aims to fast, easy and portability from high-level programming. Popular R packages handle most of low-level details and R users only focus on dataset decomposition and functional programming. Specifically, in this blog, we will show you parallel training of DNN with ‘parallel’ package and extend it to multiGPUs.

#### HOGWILD!

HOGWILD! is a data parallel model designed for stochastic gradient descent. It’s a lock-free approach with the MapReduce-like parallel-processing framework and can be used in DNN training . Thus, the training processing is designed as below:

1.Launch N workers

2.Each worker updates local weights/bias based on parts (1/N) of data

3.Master collects and averages all weights/bias from each worker

4.Each worker updates its weights/bias from master

#### Parallel in R

In R, this flow can be implemented by ‘multicores’ packages (currently, ‘parallel’ package includes both ‘multicores’ and ‘snow’ in CRAN). In below, flow chart is the standard workflow of ‘multicore’ packages with our DNN. ‘mclapply’ function creates two processors which shares memory by copy-on-write and each processor train the network by parts of data. After several steps, the master processor will do a reduce step to collect weights from two child processors and average them.In next iteration, two children will use the new weights.

Now, let’s see the details of how R code handles this data parallel model based on below real codes .

1. ‘mclapply’ creates N (devNum) workers based on ‘mc.cores’ and each worker will execute the same function, train.dnn.cublas, with different index (1:devNum);

2. the data is divided into N (devNum) parts and each worker will load their data simultaneously by their ID then the computation, even writing, can be ideally parallelized;

3. all workers exit when ‘mclapply’ is done and the results from every worker will be saved in a list (res). Master continues to remain parts and then calculate the average of all weights and bias.

4. in the next loop, the ‘mclapply’ will use the averaged model (para.model) to train again.

# Parallel Training res <-mclapply(1:devNum, function(id) { train.dnn.cublas(x, y, omodel=para.model, taindata=traindata[N.start[id]:N.end[id],], devType=“GPU”, devID=(id-1), . . .) }, mc.cores=devNum, mc.preschedule=TRUE) # Construct new model with parallel weights D <- res[[1]]$D H <- res[[1]]$H K <- res[[1]]$K for(i in 2:devNum) { res[[1]]$W1 <- res[[1]]$W1 + res[[i]]$W1 res[[1]]$W2 <- res[[1]]$W2 + res[[i]]$W2 res[[1]]$b1 <- res[[1]]$b1 + res[[i]]$b1 res[[1]]$b2 <- res[[1]]$b2 + res[[i]]$b2 } para.model <- list( D = D, H = H, K = K, # weights and bias W1= res[[1]]$W1/devNum, b1= res[[1]]$b1/devNum, W2= res[[1]]$W2/devNum, b2= res[[1]]$b2/devNum)

#### Extent to MultiGPU

Then scale to multiple GPUs, the workflow is almost as similar as CPU and only different is that each worker needs to set GPU ID explicitly and then run previous CUDA accelerated code. In other words, users still are able to access the same CUDA codes that they usually use (almost) without any change! In our implementation, we adopt the strategy of a consistent one-to-one match between CPU worker with GPU by setting GPU index as below.

// for multiple GPU case cudaSetDevice(gpuID);

#### Performance Showcase

Finally, we analyzed the performance of CPU and GPU code. The line plot shows the strong scalability of native R code (1 hidden layer and 512 neurons). And compared with H2O, native R exhibits good scalability with the number of thread increase. Look at GPU parts, one thread with one GPU is faster than 20 threads CPU implementation (both native R and H2O). Next, look at GPU scalability in bar plot where **5 times **speedup under 6 GPUs are reached and our algorithm achieved **160 times** speedup compared with original R code.

Testing on : CPU: Ivy Bridge E5-2690 v2 @ 3.00GHz, dual socket 10-core, 128G RAM; GPU: NVIDIA K40m, 12G RAM

## 3 Comments

This is the right blog for anyone who wants to learn about this matter.

You realize so much its almost hard to argue with you (not that I

really would need… HaHa). You undoubtedly set a new spin on a subject

thats been written about for years. Excellent items, simply fantastic!

You made some decent points there. I looked on the internet for the issue and found most people will go along with with your

website.

This is excellent. Great results. Most importantly such results could be achieved without expensive hardware. Thanks for the insights!

## 1 Trackback