Statistics
Mixing precision for model acceleration
A mixedprecision approach for modeling large geospatial datasets can achieve benchmark accuracy with a fraction of the computational run time.
By applying highprecision calculations only where they’re needed most, a KAUSTled research team has been able to significantly speed up modeling of large geospatial datasets without overall precision loss. The approach, implemented on a highperformance computing system based on highly parallelized graphics processor units (GPUs), will allow larger datasets to be analyzed in shorter time.
While computers have the capacity to perform very large calculations very quickly, the result can sometimes be less precise than hand calculations because of the limitation of how numbers are stored in digital systems. Standard or “single” precision numbers effectively have 6–9 actual decimal digits of precision, meaning that any calculation resulting in a longer sequence of digits will be truncated, thereby losing information.
While doubleprecision numbers can be used, this doubles the memory and calculation intensity. For geospatial datasets where the accumulation of such precision errors can lead to erroneous modeling results, this imparts a major limitation on the size of dataset that can be calculated precisely.
Sameh Abdulah and colleagues including Hatem Ltaief, Marc Genton, Ying Sun and David Keyes from KAUST, in collaboration with researchers from the University of Tennessee, Knoxville (UTK) in the U.S., have now developed an elegant solution to this problem by mixing precision as required.
“For decades, modeling of environmental data relied on doubleprecision arithmetic to predict missing data,” says Abdulah. “Today, there is highperformance computing hardware that can run single and halfprecision arithmetic with a speedup of 16 and 32 times compared with doubleprecision arithmetic. To take advantage of this, we propose a threeprecision framework that can exploit the acceleration of lower precision while maintaining accuracy by using doubleprecision arithmetic for vital information.”
Using the PaRSEC runtime system developed by UTK, which allows for ondemand precision and the orchestration of tasks and data movement across multiple parallel GPUs, the researchers exploited the statistical relationships in the data to reduce precision for weakly correlated spatial locations to single or halfprecision based on distance.
Doubleprecision calculations are only applied for the strongly correlated locations that have the most influence on model accuracy.
“The main goal of this project is to leverage the recent parallel linear algebra algorithms developed by KAUST’s Extreme Computing Research Center to scale up geospatial statistics applications on leadingedge parallel architectures,” says Abdulah.
“We have shown that we can achieve significant speedup compared to full doubleprecision arithmetic modeling while preserving the parameter estimations and prediction accuracy to meet the application requirements,” he explains. “Next, we intend to integrate approximations with mixed precision to further reduce memory footprint and shorten calculation time.”
References

Abdulah, S., Cao, Q., Pei, Y., Bosilca, G., Dongarra, J., Genton, M., Keyes, D., Ltaief, H. & Sun, Y. Accelerating geostatistical modeling and prediction with mixedprecision computations: A highproductivity approach with PaRSEC. IEEE Transactions on Parallel and Distributed Systems 2021. article
You might also like
Bioengineering
Safeguarding the right to be forgotten
Applied Mathematics and Computational Sciences
Neural networks give deeper insights
Applied Mathematics and Computational Sciences
Staying online in the high speed tube of the future
Applied Mathematics and Computational Sciences
Measuring the impact of desert greening
Statistics
Realtime modeling a step closer to reality
Bioengineering
Labgrown skin helps unlock secrets of mpox virus infection
Bioengineering
Understanding the intelligence of decision making
Applied Mathematics and Computational Sciences