Abstract
Even though deep learning shows impressive results in several applications, its use on problems with High Dimensions and Low Sample Size, such as diagnosing rare diseases, leads to overfitting. One solution often proposed is feature selection. In deep learning, along with feature selection, network sparsification is also used to improve the results when dealing with high dimensions low sample size data. However, most of the time, they are tackled as separate problems. This paper proposes a new approach that integrates feature selection, based on sparsification, into the training process of a deep neural network. This approach uses a constrained biobjective gradient descent method. It provides a set of Pareto optimal neural networks that make a trade-off between network sparsity and model accuracy. Results on both artificial and real datasets show that using a constrained biobjective gradient descent increases the network sparsity without degrading the classification performances. With the proposed approach, on an artificial dataset, the feature selection score reached 0.97 with a sparsity score of 0.92 with an accuracy of 0.9. For the same accuracy, none of the other methods reached a feature score above 0.20 and sparsity score of 0.35. Finally, statistical tests validate the results obtained on all datasets.
Publisher
Public Library of Science (PLoS)
Reference44 articles.
1. High-dimensional problems: p n;T Hastie;The Elements of Statistical Learning: Data Mining, Inference, and Prediction,2009
2. Feature selection with neural networks;A Verikas;Pattern recognition letters,2002
3. Borisov V, Haug J, Kasneci G. Cancelout: A layer for feature selection in deep neural networks. In: Artificial Neural Networks and Machine Learning–ICANN 2019: Deep Learning: 28th International Conference on Artificial Neural Networks, Munich, Germany, September 17–19, 2019, Proceedings, Part II 28. Springer; 2019. p. 72–83.
4. Liu B, Wei Y, Zhang Y, Yang Q. Deep Neural Networks for High Dimension, Low Sample Size Data. In: IJCAI; 2017. p. 2287–2293.
5. Group sparse regularization for deep neural networks;S Scardapane;Neurocomputing,2017