I have an install that has:
~1000 visitors per day
~100000 unique urls
~3 months old
DB is on a different system from Piwik.
I have turned off the automatic archiving, and am currently running the archive.sh manually at the shell.
It pegs my CPU, grids at my swap, and takes over an hour to complete if at all.
I can assume it is because I have too many unique URLs and Piwik doesn’t like that.
So I put in some JS magic to reduce the 100k urls into groups that will in time only give me 1k urls.
But I am left with a problem: How can I hack my old logs so I still have the # of visitors etc while dropping or grouping the unique urls that they went to so that the archive.sh can work normally again?
Simply put, how can I make archive.sh work properly again with the minimum of lost data?