Hi All
I'm working with a large amount of data, the largest set I've seen in real life so far was over 28k rows.
What I need to do is get it down to a compressed amount of rows based on essentially removing duplicates in a few columns, but counting certain column values into a total.
I have created this test file: VBA Test Data.xlsx
In it we have sheet "Original Data" this is essentially the raw data that I have, however its typically around 8k to 28k rows.
The sheet "Converted Data" is what it needs to compress down to.
Basically I need to do the below:
So basically just after your thoughts on how best to tackle this in the most efficient manner, as with the amount of data it can take a while to process etc.
Thanks
I'm working with a large amount of data, the largest set I've seen in real life so far was over 28k rows.
What I need to do is get it down to a compressed amount of rows based on essentially removing duplicates in a few columns, but counting certain column values into a total.
I have created this test file: VBA Test Data.xlsx
In it we have sheet "Original Data" this is essentially the raw data that I have, however its typically around 8k to 28k rows.
The sheet "Converted Data" is what it needs to compress down to.
Basically I need to do the below:
- Remove duplicate rows based on columns A / B / C / F
- However the remaining non duplicate row needs to have a count of all removed duplicates from column E & G
- End up with totals from column E & G for each unique invoice / product / code etc (like example in sheet "Converted Data"
- The original data can be in any order, and is often not in a nice grouped order like the example spreadsheet.
So basically just after your thoughts on how best to tackle this in the most efficient manner, as with the amount of data it can take a while to process etc.
Thanks