You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have a very large dataset, around 500 gigabytes, which contains various folders and subfolders with multiple Excel files. Due to limited RAM, it is not possible to use these files simultaneously (especially for data normalization). Is there a solution or example for this issue? The folders are structured as follows, with each Excel file being a multivariate time series.
With Orion, you should model each entity separately. For example, if PP17 is a different entity than PP18, then you should create two models, one for each.
For scalability, please refer to issue #567 where I suggest some solutions to loading the data in-memory.
With Orion, you should model each entity separately. For example, if PP17 is a different entity than PP18, then you should create two models, one for each.
For scalability, please refer to issue #567 where I suggest some solutions to loading the data in-memory.
Thank you for the clarification and the helpful reference. These files actually correspond to different sensors, but they are present in all files. Our goal is to train the network on this normal operational data and then test it on faulty data. Are you suggesting that we need to develop a separate model for each sensor, rather than using an approach like LSTM AE where multiple sensors are fees to model, and the reconstruction error is measured?
Yes @jafa2222, current models in Orion support one-sensor detections rather than multiple. This is because most models are learn a pattern of one signal, making it better at learning what the expected pattern should look like.
The first blog post we highlight in our readme talks about the general framework followed by these models if you'd like to read more about this topic!
Description
I have a very large dataset, around 500 gigabytes, which contains various folders and subfolders with multiple Excel files. Due to limited RAM, it is not possible to use these files simultaneously (especially for data normalization). Is there a solution or example for this issue? The folders are structured as follows, with each Excel file being a multivariate time series.
The text was updated successfully, but these errors were encountered: