How to avoid memory problem while processing huge table?
Mostrar comentarios más antiguos
I have a huge observation table with around 30 Lacs of rows and 12 columns. While training knn classifier in 2016a version, I am getting errors related to memory. Is there any way to avoid this? I have tried to reduce rows but it's affecting the output quality.
Each row in table is a pixel and it's other values as features in columns. In one set of MRI scan, there are around 20 images of 512x512, I am loading one set for creating observation table. Is there another way to pass large amount of data to knn classifier?
Respuestas (1)
KSSV
el 31 de Ag. de 2016
1 voto
doc datastore, memmap, mapreduce.
1 comentario
Nitinkumar Ambekar
el 1 de Sept. de 2016
Categorías
Más información sobre Image Segmentation en Centro de ayuda y File Exchange.
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!