I am doing my own visitor tracking with special features that Google Analytics (nor any other) can provide me as it is customized. I was calling this function near the end of my script, but quickly ran into our clients running into thousands of pages being called from bots (I assume Google), and my table filled up with around 1,000,000 useless and deceptive records in the period of a month.
The method most people use is they use Javascript at the bottom of the page, bots don't operate javascript and so this is an easy fix -- but I am looking for PHP solution.
The last thing I did was use PHP's get_browser
:
http://us2.php.net/manual/en/function.get-browser.php
and check for the crawler
aspect. This didn't work.
I have looked at this post: Stopping scripters from slamming your website hundreds of times a second
But the main solution to that was doing something similar to SO where it brings up a CAPTCHA. My point is not to stop the bots -- I want the pages crawled. I simply don't want to send my visitor tracking data when they are there.
I switched to Javascript solution right now, performing an AJAX request, as our users were getting irritated and the inaccurate statistics.