Optimal Method For Saving Large Amounts of Data To Read Into MATLAB

2 vues (au cours des 30 derniers jours)
Cole Smith
Cole Smith le 20 Fév 2015
Commenté : Tejash Patel le 25 Juin 2018
Hi, I have somewhat of a unique problem in that I have a box with a constant stream of time-series network data being received and filtered, that I eventually want to read into MATLAB. This data can be several gigabytes per hour, and my MATLAB application only needs to query specific messages/timeframes. I know MATLAB uses HDF5 internally, so that looked like a good option but there's really no time-series capabilities. Other wrappers such as PyTables/TsTables exist to add time series to HDF5, but that requires an extra python wrapper. I've also looked into things like InfluxDB, but that doesn't save in the native HDF5 format, so I think there's a good chance large queries could be extremely slow if executed in MATLAB.
Has anyone done something similar or know of a good way to store continuous network data that I want to search/open in MATLAB at a later time?

Réponses (1)

Greig
Greig le 22 Fév 2015
While I'm sorry to say that I don't have a direct solution for you, perhaps you will get some useful ideas from these blogs on handling big data in MATLAB...

Catégories

En savoir plus sur Deep Learning Toolbox dans Help Center et File Exchange

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by