Matlab unable to parse a Numeric field when I use the gather function on a tall array.
14 visualizaciones (últimos 30 días)
Mostrar comentarios más antiguos
Ninad
el 21 de Ag. de 2025
Comentada: Jeremy Hughes
el 25 de Sept. de 2025
So I have a CSV file with a large amount of datapoints that I want to perform a particular algorithm on. So I created a tall array from the file and wanted to import a small chunk of the data at a time. However, when I tried to use gather to get the small chunk into the memory, I get the following error.

"Board_Ai0" is the header of the CSV file. It is not in present in row 15355 as can be seen below where I opened the csv file in MATLAB's import tool.

The same algorithm works perfectly fine when I don't use tall array but instead import the whole file into the memory. However, I have other larger CSV files that I also want to analyze but won't fit in memory.
UPDATE: So apparently the images were illegible but someone else edited the question to make the size of the image larger so I guess it should be fine now. Also I can't attach the data files to this question because the data files that give me this problems are all larger than 5 GB.
12 comentarios
Jeremy Hughes
el 25 de Sept. de 2025
FYI, a TALL array is meant to allow you to operate on the entire table, even if it doesn't fit into memory. If you want to work on chunks of the file, don't use TALL.
Using datastore directly will let you read chunks of data.
ds = tabularTextDatastore(files,....)
while hasdata(ds)
T = read(ds)
% Do stuff.
end
However, that's not going to solve the problem because you have rows that don't contain numeric data. tabularTextDatastore doesn't allow for that.
I like @Harald's solution--but with some modification. I'd avoid calling detectImportOptions every iteration. For a datastore to work, the schema should be the same each time.
function [data,startrow,done] = readdata(filename,startrow)
persistent opts
if isempty(opts)
opts = detectImportOptions(filename);
end
nRows = 10000000;
if isempty(startrow)
startrow = 2;
end
opts.DataLines = [startrow, startrow+nRows-1];
data = readtimetable(filename, opts);
data = rmmissing(data);
done = height(data) < nRows;
startrow = startrow + nRows;
end
There is still a problem with this; using opts.DataLines to manage the chunks still forces you to read lines up to the startRow in order to know where to start. That will mean each subsequent read will be slower.
Respuesta aceptada
Daniele Sportillo
el 12 de Sept. de 2025
Hi @Ninad, thanks for sharing the file. I see that your .csv includes the Variable Names in some data rows.


To handle this, you can use the TreatAsMissing property with tabularTextDatastore to treat those rows as NaN
data = tabularTextDatastore("1kwogndrd1.csv",TreatAsMissing={'Time','Board0_Ai0'},SelectedVariableNames={'Board0_Ai0'});
Then you can gather slices of the data without errors:
ds = tall(data);
slice = ds.Board0_Ai0(1:200000);
slice = gather(slice);
If you want to calculate the mean on the entire column, I recommend computing it first and then gathering the result. Use "omitnan" to exclude the NaN rows introduced by TreatAsMissing
m = mean(ds.Board0_Ai0,"omitnan");
gather(m)
Hope this helps!
1 comentario
Jeremy Hughes
el 25 de Sept. de 2025
I"d just use the datastore directly instead of using TALL. Assuming you want the mean of each chunk.
ds = tabularTextDatastore("1kwogndrd1.csv",TreatAsMissing={'Time','Board0_Ai0'},SelectedVariableNames={'Board0_Ai0'});
M = {};
while hasdata(ds)
data = read(ds);
M{end+1} = mean(data)
end
If you want the mean of the entire variable, TALL doesn't need to be chunked.
ds = tabularTextDatastore("1kwogndrd1.csv",TreatAsMissing={'Time','Board0_Ai0'},SelectedVariableNames={'Board0_Ai0'});
data = tall(ds)
M = mean(data)
M = gather(M)
Más respuestas (2)
dpb
el 6 de Sept. de 2025
Editada: dpb
el 6 de Sept. de 2025
It appears it is detectImportOptions that is having the problem -- apparently it tries to read the whole file into memory first before it does its forensics.
I don't think you need an import options object anyway, use the 'Range' named parameter in the argument to readtimetable
Something like
function [data,startrow,done] = readdata(filename,startrow)
nRows = 10000000;
if isempty(startrow)
startrow = 2; % this looks unlikely to be right from the earlier image there are 3(?) header rows?
end
range=sprintf('%d:%d',startrow, startrow+nRows); % build row range expression
data = readtimetable(filename, 'Range',range);
data = rmmissing(data);
done = height(data) < nRows;
startrow = startrow + nRows;
end
This may still have some issues using the timetable, however if it first reads variable names from a header line which header line isn't there in the subsequent sections of the file. I don't know what trouble you'll run into with such large files if try to read 100K lines into the file but tell it to also read the variablenames from the second or third line in the file....probably ignoring variable names and letting MATLAB use defaults then set the Properties.VariableNames after reading of just accept the defaults would be best bet.
5 comentarios
Harald
el 9 de Sept. de 2025
dpb
el 9 de Sept. de 2025
Editada: dpb
el 9 de Sept. de 2025
@Harald, no problem, just commenting on why I hadn't poked harder, earlier...
If @Ninad would attach a short section of a file it would make it simpler, indeed. It's not convenient at the moment to stop a debugging session and try to create a local copy of a similar file to play with/poke at.
The documentation isn't all that helpful, the only examples I can find using tables/timetables with tall arrays are tiny data files and don't use the filedatastore so they don't have a callback function with a table. I don't believe there is an example of the combination....
Stephen23
el 8 de Sept. de 2025
Editada: dpb
el 8 de Sept. de 2025
Providing the RANGE argument does not prevent READTABLE from calling its automatic format detection:
which might involve loading all or a significant part of the file into memory. The documented solution is to provide an import options object yourself (e.g. you can generate this on a known good file of a smaller size and then storing it) or alternatively using a low-level file reading command, e.g. FSCANF, FREAD, etc.
4 comentarios
dpb
el 11 de Sept. de 2025
I was suggesting to attach a piece of the file (perhaps zipped to include a little more). That would be enough for folks to have enough to test with that duplicates the actual format.
What, precisely, does "MATLAB crashed" mean? Actually aborted MATLAB itself or another out-of-memory or ...?
Ver también
Categorías
Más información sobre Large Files and Big Data en Help Center y File Exchange.
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!




