/ akbir khan / blog

# golang's rand and concurrency

April 27, 2020

This is a bug I encountered when writing a ray-tracer in Golang and thought it was worth a quick share.

### The Idea

When rendering an image using a ray-tracer, each pixel value can theoretically be calculated independently, making parallelisation an attractive prospect. When rendering the image, we iterate through each position and call the sample method to calculate RGB value for that pixel:

Go makes running threads (go routines) really simple and so the idea was to split the for loop into separate threads. Where the k-th thread would calculate the [k, 2k, ... , nk]-th position pixels within the row.

Now lets check the runtimes:

# Single Thread
Elapsed: 1m54.230536119s
Elapsed: 4m13.680876913s


Unfortunately this didn’t speed things up, even worse, it increased the runtime!

### The Bug

If threading was making my program slower, then each thread was clearly unable to act independently. So lets check out sample method:

After some debugging (shout out pprof), I noticed that an incredible amount of time was being used by math/rand and the associated synchronisation overhead. Taking a quick look at the source code, I notice that the package is globally locked:

For random number generators (RNGs) this makes perfect sense, as RNGs generally calculate the next random number based on the previous, and we’d like RNG to be consistently random across processes. This is good behaviour — however not ideal in our situation.

### The Fix

To circumvent this issue and gain all the speed benefits of threading, we’d need to provide each thread its own RNG. Thankfully this is fairly simple within go, as the rand package has methods to create unique RNGs, which can be created for each thread independently.

and subsequently update all the existing math/rand calls must be switched to call our RNG instead:

### The Results

After some tuning of the number of threads to use (as suggested by a friend 2*NumCores), my final speed up was:

# Multiple Threads
Elapsed: 23.765684225s


which was a nice 4x speed up!