Please use this identifier to cite or link to this item:
|Title:||An improved CUDA-based implementation of differential evolution on GPU||Authors:||Raimondo, Federico
Ong, Yew Soon
Qin, A. K.
|Keywords:||DRNTU::Engineering::Computer science and engineering||Issue Date:||2012||Source:||Qin, A. K., Raimondo, F., Forbes, F., & Ong, Y. S. (2012). An improved CUDA-based implementation of differential evolution on GPU. Proceedings of the fourteenth international conference on Genetic and evolutionary computation conference, 991-998.||Abstract:||Modern GPUs enable widely affordable personal computers to carry out massively parallel computation tasks. NVIDIA's CUDA technology provides a wieldy parallel computing platform. Many state-of-the-art algorithms arising from different fields have been redesigned based on CUDA to achieve computational speedup. Differential evolution (DE), as a very promising evolutionary algorithm, is highly suitable for parallelization owing to its data-parallel algorithmic structure. However, most existing CUDA-based DE implementations suffer from excessive low-throughput memory access and less efficient device utilization. This work presents an improved CUDA-based DE to optimize memory and device utilization: several logically-related kernels are combined into one composite kernel to reduce global memory access; kernel execution configuration parameters are automatically determined to maximize device occupancy; streams are employed to enable concurrent kernel execution to maximize device utilization. Experimental results on several numerical problems demonstrate superior computational time efficiency of the proposed method over two recent CUDA-based DE and the sequential DE across varying problem dimensions and algorithmic population sizes.||URI:||https://hdl.handle.net/10356/100559
|Appears in Collections:||SCSE Conference Papers|
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.