Large Excel Files (100 Mb) - Extracting partial information #431

Closed
opened 2016-06-27 22:33:00 +00:00 by amerj19 · 4 comments
amerj19 commented 2016-06-27 22:33:00 +00:00 (Migrated from github.com)

Hi

Can I read large excel files (around 100 MB)? I tried with the sample but browser crashes. I understand, I can read file in slices, however, I am not sure how to read information how to read data from each slice. e.g. cell values etc. Is there a good example to read large excel files on client.

Hi Can I read large excel files (around 100 MB)? I tried with the sample but browser crashes. I understand, I can read file in slices, however, I am not sure how to read information how to read data from each slice. e.g. cell values etc. Is there a good example to read large excel files on client.
fabriziomorello commented 2016-06-30 15:42:14 +00:00 (Migrated from github.com)

I have the same issue. Hope someone can help us.

I have the same issue. Hope someone can help us.
diegoles commented 2016-08-03 20:30:58 +00:00 (Migrated from github.com)

I have the same issue. Hope someone can help us.

I have the same issue. Hope someone can help us.
reviewher commented 2017-01-15 00:01:57 +00:00 (Migrated from github.com)

@amerj19 @fabriziomorello @diegoles I recommend trying to use the node-based tool first to make sure that the files can be read and that the issue is due to the size. Alternatively you can email us or post a link to a file that crashes the browser.

As for the general question, the way the excel data is stored, you need to process quite a bit. The container formats themselves (zip and cfb) are so flexible that a reader must load all of the data in memory (some write tools scatter metadata throughout the file). In the XLSX case, you additionally have to read multiple sub-files to find the cell values. So there's no obvious way to reduce memory consumption without radically changing the API surface.

@amerj19 @fabriziomorello @diegoles I recommend trying to use the node-based tool first to make sure that the files can be read and that the issue is due to the size. Alternatively you can email us or post a link to a file that crashes the browser. As for the general question, the way the excel data is stored, you need to process quite a bit. The container formats themselves (zip and cfb) are so flexible that a reader must load all of the data in memory (some write tools scatter metadata throughout the file). In the XLSX case, you additionally have to read multiple sub-files to find the cell values. So there's no obvious way to reduce memory consumption without radically changing the API surface.
reviewher commented 2021-09-19 22:40:58 +00:00 (Migrated from github.com)

Combining all of these discussions about reading large files into one issue: #61

Combining all of these discussions about reading large files into one issue: #61
Sign in to join this conversation.
No Milestone
No Assignees
1 Participants
Notifications
Due Date
The due date is invalid or out of range. Please use the format 'yyyy-mm-dd'.

No due date set.

Dependencies

No dependencies set.

Reference: sheetjs/sheetjs#431
No description provided.