Ahora está siguiendo esta publicación
- Verá actualizaciones en las notificaciones de contenido en seguimiento.
- Podrá recibir correos electrónicos, en función de las preferencias de comunicación que haya establecido.
A new meta-heuristic algorithm called the Mountain Gazelle Optimizer (MGO) was developed in part as a result of wild mountain gazelles' social structure but suffered from slow convergence speed. Consequently, a modified MGO (mMGO) approach uses the Joint Opposite Selection (JOS) operator, which combines the Selective Leading Opposition (SLO) and the Dynamic Opposite Learning (DO) approaches, to improve MGO. The purpose of this study is to evaluate the performance of mMGO based on the k-Nearest Neighbor (kNN) classifier in predicting brain stroke in data sets taken from Kaggle. Performance was assessed on the challenging CEC 2020 benchmark test functions. Compared to seven well-known optimization algorithms, the statistical results demonstrated the superiority of mMGO. Furthermore, the experimental results of mMGO-kNN for categorizing brain stroke data sets revealed that it outperformed competitors in all data sets with an overall accuracy of 95.5\%, a sensitivity of 99.34\%, a specificity of 98.99\%, and a precision of 99.21\%.
Citar como
Prof. Dr. Essam H Houssein (2026). mMGO for Brain Stroke Classification (https://la.mathworks.com/matlabcentral/fileexchange/157441-mmgo-for-brain-stroke-classification), MATLAB Central File Exchange. Recuperado .
Agradecimientos
Inspirado por: Feature Selection, classification k-means, Global Optimization with MATLAB
Información general
- Versión 1.0.0 (872 KB)
Compatibilidad con la versión de MATLAB
- Compatible con cualquier versión
Compatibilidad con las plataformas
- Windows
- macOS
- Linux
| Versión | Publicado | Notas de la versión | Action |
|---|---|---|---|
| 1.0.0 |
