import_logs.py - avoid tracking some pages like robots.txt

hi there,

I’m using import_logs.py and would like to put certain pages on blacklist to be actually ignored during import.

e.g. I have few clients everyday accesing robots.txt which have normal user agent nor they contain *bot or whatever I could filter for (neither by IP). - is there a way how to exclude some given paths/pages from log import?

thanks & cheers

Yes you can exclude specific paths with:


--exclude-path=EXCLUDED_PATHS
                        Paths to exclude. Can be specified multiple times
  --exclude-path-from=EXCLUDE_PATH_FROM
                        Each line from this file is a path to exclude

Thanks, works fine.

cheers