A new meta-heuristic algorithm called the Mountain Gazelle Optimizer (MGO) was developed in part as a result of wild mountain gazelles' social structure but suffered from slow convergence speed. Consequently, a modified MGO (mMGO) approach uses the Joint Opposite Selection (JOS) operator, which combines the Selective Leading Opposition (SLO) and the Dynamic Opposite Learning (DO) approaches, to improve MGO. The purpose of this study is to evaluate the performance of mMGO based on the k-Nearest Neighbor (kNN) classifier in predicting brain stroke in data sets taken from Kaggle. Performance was assessed on the challenging CEC 2020 benchmark test functions. Compared to seven well-known optimization algorithms, the statistical results demonstrated the superiority of mMGO. Furthermore, the experimental results of mMGO-kNN for categorizing brain stroke data sets revealed that it outperformed competitors in all data sets with an overall accuracy of 95.5\%, a sensitivity of 99.34\%, a specificity of 98.99\%, and a precision of 99.21\%.
引用格式
Prof. Dr. Essam H Houssein (2024). mMGO for Brain Stroke Classification (https://www.mathworks.com/matlabcentral/fileexchange/157441-mmgo-for-brain-stroke-classification), MATLAB Central File Exchange. 检索时间: .
MATLAB 版本兼容性
创建方式
R2023b
兼容任何版本
平台兼容性
Windows macOS Linux标签
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!mMGO with Datasets
mMGO with Datasets/HGS
mMGO with Datasets/HHO
mMGO with Datasets/MFO
mMGO with Datasets/MGO
mMGO with Datasets/WOA
版本 | 已发布 | 发行说明 | |
---|---|---|---|
1.0.0 |