- avoid tracking some pages like robots.txt

hi there,

I’m using and would like to put certain pages on blacklist to be actually ignored during import.

e.g. I have few clients everyday accesing robots.txt which have normal user agent nor they contain *bot or whatever I could filter for (neither by IP). - is there a way how to exclude some given paths/pages from log import?

thanks & cheers

Yes you can exclude specific paths with:

                        Paths to exclude. Can be specified multiple times
                        Each line from this file is a path to exclude

Thanks, works fine.