Author:
Perez-Wohlfeil Esteban,Trelles Oswaldo,Guil Nicolás
Abstract
AbstractThe use of Graphics Processing Units to accelerate computational applications is increasingly being adopted due to its affordability, flexibility and performance. However, achieving top performance comes at the price of restricted data-parallelism models. In the case of sequence alignment, most GPU-based approaches focus on accelerating the Smith-Waterman dynamic programming algorithm due to its regularity. Nevertheless, because of its quadratic complexity, it becomes impractical when comparing long sequences, and therefore heuristic methods are required to reduce the search space. We present GPUGECKO, a CUDA implementation for the sequential, seed-and-extend sequence-comparison algorithm, GECKO. Our proposal includes optimized kernels based on collective operations capable of producing arbitrarily long alignments while dealing with heterogeneous and unpredictable load. Contrary to other state-of-the-art methods, GPUGECKO employs a batching mechanism that prevents memory exhaustion by not requiring to fit all alignments at once into the device memory, therefore enabling to run massive comparisons exhaustively with improved sensitivity while also providing up to 6x average speedup w.r.t. the CUDA acceleration of BLASTN.
Funder
European project ELIXIR-EXCELERATE
Spanish national project Plataforma de Recursos Biomoleculares y Bioinformáticos
Fondo Europeo de Desarrollo Regional
Instituto de Investigación Biomédica de Málaga
University of Málaga
Junta de Andalucía
Universidad de Málaga
Publisher
Springer Science and Business Media LLC
Subject
Hardware and Architecture,Information Systems,Theoretical Computer Science,Software
Reference48 articles.
1. Owens JD, Luebke D, Govindaraju N, Harris M, Krüger J, Lefohn AE, Purcell TJ: A survey of general-purpose computation on graphics hardware. In: Computer Graphics Forum, vol. 26, pp. 80–113 (2007). Wiley Online Library
2. Navarro CA, Hitschfeld-Kahler N, Mateu L (2014) A survey on parallel computing and its applications in data-parallel problems using gpu architectures. Commun Comput Phys 15(2):285–329
3. Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J, Devin M, Ghemawat S, Irving G, Isard M et al. (2016) Tensorflow: A system for large-scale machine learning. In: 12th symposium on operating systems design and implementation, vol 16, pp 265–283
4. Stone JE, Hardy DJ, Ufimtsev IS, Schulten K (2010) Gpu-accelerated molecular modeling coming of age. J Mol Gr Modell 29(2):116–125
5. Lu F, Song J, Cao X, Zhu X (2012) Cpu/gpu computing for long-wave radiation physics on large gpu clusters. Computers Geosci 41:47–55
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献