Tag Archives: Parallelization

Colfax Hands-On Webinar Series: Deep Dive into Performance Optimization

Free 20-hour webinar series includes parallel programming, performance optimization, remote access to advanced servers Intel partner Colfax Research is offering a free 20-hour hands-on in-depth training on parallel programming and performance optimization in computational applications on Intel architecture. The first run in 2017 begins January 16, 2017. Broadcasts start at 17:00 UTC (9:00 am in […]

Posted in News, Tune | Tagged , , | Leave a comment

Data Races: What They Are, How to Fix Them

I have talked a lot about the parallelization of loops using OpenMP. It is an easy way to improve performance in your applications, especially if you can apply the technique to loops that happen often or loops with many iterations. In many cases, OpenMP provides optimized performance with no down-side risks. But there are other […]

Posted in Tune, Uncategorized | Tagged , | Leave a comment

What is the Effect of Simultaneous OpenMP Loops?

OpenMP simplifies code parallelization, but can one overdo their use of this valuable tool? In the blog Slashdot Media Contributing Editor Rick Leinecker creates some gnarly code to see if it creates a performance hit I have spent a lot of time here at Go Parallel talking about OpenMP loops. The OpenMP standard provides simple […]

Posted in Tune | Tagged , | Leave a comment

Breaking Down OpenMP Loops

OpenMP can bring amazing performance boosts to your applications. This presentation breaks down OpenMP loops that have no dependencies. It also shows how easy it is to parallelize with OpenMP by using compiler directives.

Posted in Tune, Video | Tagged , | Leave a comment

Improving Data Compression: a Parallel Algorithm for “Shannon Entropy”

A great deal of my personal research is in the area of data compression. I have been doing this type of research for about 20 years. A closely-related topic is data entropy. Data entropy is similar to the thermodynamic entropy that many people think of. The higher the data entropy, the more chaotic and unpredictable […]

Posted in Tune | Tagged , | 1 Comment

Threading Building Blocks (TBB) at 10: Intel’s James Reinders

What is the impact of Intel’s Threaded Building Blocks for Multiprocessing? Listen to this interview with James Reinders to recap 10 years of TBB Author and parallel computing James Reinders recently retired after a brilliant 25 year career at Intel.  Just prior to his retirement, James sat down with Intersect 360’s Addison Snell to discuss […]

Posted in Build | Tagged , , | Leave a comment

OpenMP: Past the Basics

So, you’ve started tinkering with OpenMP to help parallelize your code. Now what?  This video by Slashdot Media Contributing Editor Rick Leinecker points OpenMP newbies in the right direction to go beyond parallelizing for loops, and demonstrates how to avoid data race conditions while you’re doing it.  

Posted in Tune, Video | Tagged , , | Leave a comment

Parallelizing Binary Searches

Binary searches are orders of magnitude faster than linear searches. In this tutorial, Slashdot Media Contributing Editor Rick Leinecker shows how to speed all manner of searches in your code by paralellizing using OpenMP

Posted in Tune | Tagged , | Leave a comment

Avoiding Data Races with Reducers

Fixing what happens when simultaneous threads modify the SAME memory.

Parallelizing code drastically speeds execution, but if you’re not careful data race conditions can produce very unwanted results. See how Slashdot Media Contributing Editor Rick Leinecker gets at the problem and fixes it automatically with a reduction clause.

Posted in Tune | Tagged , | Leave a comment

Timing C++ Execution: Do Compilers Matter?

Execution speed can be impacted by choice of language AND compiler In this instructional video, Slashdot Media Contributing Editor David Bolton show us how to get the most performance out of C++ code by paying attention to compiler features that are in tune with hardware features.

Posted in Tune | Tagged , | 1 Comment