Variable Selection for Recurrent Events Using Heuristic Approaches: Identifying Informative Variables for Rehospitalization in Schizophrenia Patients
Abstract
Introduction: Recurrent event data, as a generalization of survival data, are frequently observed in various areas of medical research, including sequential hospitalizations in patients with schizophrenia. As experiencing multiple relapses during schizophrenia can have many implications, such as self-harm or harm to others, loss of education or employment, or other adverse outcomes, identifying and determining the most critical factors related to relapses in this disorder is essential. This study aimed to utilize heuristic approaches for selecting predictor variables in the field of recurrent events with an application to schizophrenia disorder.
Methods: A two-step algorithm was employed to apply a combination of two variable selection methods, recursive feature elimination (RFE) and genetic algorithm feature selection (GAFS), and four modeling techniques: Gradient boosting (GB), artificial neural network (ANN), random forest (RF), and support vector machine (SVM) to simulated recurrent event datasets.
Results: In most simulation scenarios, the results indicated that the combination of RFE and RF applied to the deviance residual (DR) outperforms the other methods. The RFE-RF-DR selected the following predictor variables: Number of children, age, marital status, and history of substance abuse.
Conclusions: Our findings revealed that the proposed machine learning-based model is a promising technique for selecting predictor variables associated with a recurrent outcome when analyzing multivariate time-to-event data with recurrent events.
2. Cook RJ, Lawless JF. The statistical analysis of recurrent events: Springer; 2007.
3. Emsley R, Chiliza B, Asmal L, Harvey BH. The nature of relapse in schizophrenia. BMC psychiatry. 2013;13(1):1-8.
4. Heinze G, Wallisch C, Dunkler D. Variable selection–a review and recommendations for the practicing statistician. Biometrical journal. 2018;60(3):431-49.
5. Hocking RR. A Biometrics invited paper. The analysis and selection of variables in linear regression. Biometrics. 1976;32(1):1-49.
6. Akaike H. A new look at the statistical model identification. IEEE transactions on automatic control. 1974;19(6):716-23.
7. Schwarz G. Estimating the dimension of a model. The annals of statistics. 1978;6(2):461-4.
8. Tibshirani R. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society: Series B (Methodological). 1996;58(1):267-88.
9. Zou H. The adaptive lasso and its oracle properties. Journal of the American statistical association. 2006;101(476):1418-29.
10. Fan J, Li R. Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American statistical Association. 2001;96(456):1348-60.
11. Zhang C-H. Nearly unbiased variable selection under minimax concave penalty. The Annals of statistics. 2010;38(2):894-942.
12. Wei P, Lu Z, Song J. Variable importance analysis: a comprehensive review. Reliability Engineering & System Safety. 2015;142:399-432.
13. Malley JD, Malley KG, Pajevic S. Statistical learning for biomedical data: Cambridge University Press; 2011.
14. Duan R, Fu H. Estimate variable importance for recurrent event outcomes with an application to identify hypoglycemia risk factors. Statistics in medicine. 2015;34(19):2743-54.
15. Therneau TM, Grambsch PM, Fleming TR. Martingale-based residuals for survival models. Biometrika. 1990;77(1):147-60.
16. Box-Steffensmeier JM, Jones BS. Event history modeling: A guide for social scientists: Cambridge University Press; 2004.
17. Cortes C, Vapnik V. Support-vector networks. Machine learning. 1995;20(3):273-97.
18. Romero E, Toppo D. Comparing support vector machines and feedforward neural networks with similar hidden-layer weights. IEEE Transactions on Neural Networks. 2007;18(3):959-63.
19. Platt J. Sequential minimal optimization: A fast algorithm for training support vector machines. 1998.
20. Theobald O. Machine learning for absolute beginners: Scatterplot press; 2017.
21. McCulloch WS, Pitts W. A logical calculus of the ideas immanent in nervous activity. The bulletin of mathematical biophysics. 1943;5(4):115-33.
22. Xia J-H, Kumta AS. Feedforward neural network trained by BFGS algorithm for modeling plasma etching of silicon carbide. IEEE transactions on plasma science. 2010;38(2):142-8.
23. Friedman JH. Greedy function approximation: a gradient boosting machine. Annals of statistics. 2001;29(5):1189-232.
24. Guelman L. Gradient boosting trees for auto insurance loss cost modeling and prediction. Expert Systems with Applications. 2012;39(3):3659-67.
25. Hastie T, Tibshirani R, Friedman J. The elements of statistical learning: Springer series in statistics; 2009.
26. Breiman L. Random forests. Machine learning. 2001;45(1):5-32.
27. Ho TK, editor Random decision forests. Proceedings of 3rd international conference on document analysis and recognition; 1995: IEEE.
28. Sandhu AK, Batth RS. Software reuse analytics using integrated random forest and gradient boosting machine learning algorithm. Software: Practice and Experience. 2021;51(4):735-47.
29. Lin C-H, Chen M-C, Chou L-S, Lin C-H, Chen C-C, Lane H-Y. Time to rehospitalization in patients with major depression vs. those with schizophrenia or bipolar I disorder in a public psychiatric hospital. Psychiatry research. 2010;180(2-3):74-9.
30. Husted J, Jorgens A. Best practices: population density as a factor in the rehospitalization of persons with serious and persistent mental illness. Psychiatric Services. 2000;51(5):603-5.
Files | ||
Issue | Vol 9 No 1 (2023) | |
Section | Original Article(s) | |
DOI | https://doi.org/10.18502/jbe.v9i1.13979 | |
Keywords | ||
machine learning; gradient boosting; deviance; recurrent event data; variable importance; schizophrenia |
Rights and permissions | |
This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License. |