So I have a project where is receive 1500+ .csv sheets of data. Each file size is about 12kb and contains one column with timestamps, and one column with data that i need to extract based on the timestamp. Each sheet is sorted and grouped (these are power readings from rack PDUs in a data center, each sheet represents 1 PDU and each rack has 4 PDUs, they need to be grouped by rack). I breakup the 1500 sheets into chunks of around 100 for workability, and have 15 of these groups.
I process each individual sheet as a separate tab in one Group Workbook. I then have a Template file that simply references external link to the Group Workbook (i do this to sort/associate each PDU to the racks and this will be an ongoing thing for the immediate future). The file size for each Group set is around 6MB, but after it is processed and pulling data from external links, each of the Template files comes out to 100MB+, some are 200+ and can take a few minutes to open and save.
My process is clunky and inefficient, but it works. However, I will likely be processing more data in the future and expect larger files sizes and longer loading times. I'm sure there is a more feasible option that I am missing. It's becoming unworkable and I need some help to make it better. Thanks
Raw data that I receive (individual rPDU readings) ~12kb
I group them by location and process associated sheets as separate tabs in an index workbook. ~600kb
I then use a Template to sort/group is reading and reference links to the Index Workbook. This puts all the PDUs together sorted by rack. Then I total the readings, and get Peak, Avg, and Min. This comes out to 100MB+ file size.
I process each individual sheet as a separate tab in one Group Workbook. I then have a Template file that simply references external link to the Group Workbook (i do this to sort/associate each PDU to the racks and this will be an ongoing thing for the immediate future). The file size for each Group set is around 6MB, but after it is processed and pulling data from external links, each of the Template files comes out to 100MB+, some are 200+ and can take a few minutes to open and save.
My process is clunky and inefficient, but it works. However, I will likely be processing more data in the future and expect larger files sizes and longer loading times. I'm sure there is a more feasible option that I am missing. It's becoming unworkable and I need some help to make it better. Thanks
Raw data that I receive (individual rPDU readings) ~12kb
I group them by location and process associated sheets as separate tabs in an index workbook. ~600kb
I then use a Template to sort/group is reading and reference links to the Index Workbook. This puts all the PDUs together sorted by rack. Then I total the readings, and get Peak, Avg, and Min. This comes out to 100MB+ file size.