Programming difficulty is killing engineers' productivity

While supercomputers are thought to help accelerate engineering and scientific discovery, a new study sponsored by Interactive Supercomputing suggests difficulties in programming is increasingly becoming one of researchers’ biggest productivity killers.

The Development of Custom Parallel Computing Applications study conducted by the Simon Management Group surveyed more than 500 users of parallel high-performance computers (HPCs) from a range of industries including education, government, aerospace, healthcare, manufacturing, geo-sciences, bio-sciences and semiconductor. The report examines the software tools currently used, probes current application development environments, practices, and limitations, and catalogues critical issues and bottlenecks.

The study indicates that parallel code writing, programming efficiency, translation, debugging and limits of HPC software are the most frequently cited bottlenecks across all industries. Respondents indicated there is an urgent need to shorten the application development time of custom algorithms and models.

The largest category of respondents (42.3 per cent) said that a typical project takes six months to complete, yet nearly 20 percent of respondents’ projects consume two to three years of their time.

The majority of parallel application prototypes (65 per cent) are developed in very high-level languages (VHLLs) such as Matlab, Mathematica, Python, and R. While C and Fortran are frequently used to prototype, respondents overwhelmingly said they would prefer to work with an interactive desktop tool if the prototype could be easily bridged to work with HPC servers.

The disconnect stems from the fact that desktop computers cannot handle the processing and memory requirements of the huge amounts of data that many scientific and engineering problems analyse. The problem is only getting worse; according to the study, the average median-sized data set used in a technical computing application today ranges from 10 to 45Gb and is expected to swell to 200 to 600Gb in just three years.

'This study demonstrates that programming tools have not kept pace with the advances in the computing hardware and affordability of high-performance computers,' said Peter Simon, president of Simon Management Group.

Twitter icon
Google icon icon
Digg icon
LinkedIn icon
Reddit icon
e-mail icon

Robert Roe looks at recent developments in processor and accelerator technology


Robert Roe discusses the potential for software-defined storage with Excelero’s Josh Goldenhar


With innovation in cooling technology increasingly more important to ensure energy, performance and cost efficiency of HPC, Keely Portway speaks to experts to find out what is driving the latest innovations


Robert Roe talks to Southampton University’s Oz Parchment about the decision-making behind the latest HPC system at the University


Gemma Church explains the background behind explosive growth in the simulation and modelling of low- and high-frequency electronics