Reading large excel file using "read" command takes ~5 minutes, is this expected performance?

1 次查看(过去 30 天)
I am reading a simulation/test output data in a .xlsx file into a matlab table through a datastore variable. The test data contains 450+ variables, each with 20000+ samples (i.e.) 450+ columns and 20000+ rows but all are numbers. I created a datastore on the excel file, modified the selected variables and variable type properties and used read command to read the file into a matlab table, it took about ~5 minutes. When I tried readtable command on the excel file directly, it took about the same time as well. However when I tried reading the file interactively using matlab export dialog, it took less than 30 seconds, so I am wondering if there's any way to achieve the same level of efficiency programmatically?

采纳的回答

J. Alex Lee
J. Alex Lee 2020-9-6
Try manually creating the import options with spreadsheetimportoptions().
  2 个评论
Ajay Kumar
Ajay Kumar 2020-9-7
Thanks. I will read up on this function and try it out. The format of the test output sheet will not change that often, but If it does, then I will have to update this object that I am gonna create?

请先登录,再进行评论。

更多回答(0 个)

类别

Help CenterFile Exchange 中查找有关 Data Import from MATLAB 的更多信息

产品


版本

R2017b

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by