T
Tuxedo
I would like to implement a web page counter to measure 'unique' visitors
p/page and p/IP in 24 hour cycles, so if several visits to a particular
page comes from a same IP in a 24 hours time frame, it counts as one
visitor to that page. While if the same visitor comes back to the same
page, say 25 hours later counting from his previous IP and page count, the
count can increment for that page again. In reality, this may represent a
repeat visitor or another visitor having been assigned the same IP.
Depending on type of web traffic and overall rotating IPs cycles by
different net connection service providers, this type of solution may at
best provide a highly approximate overview of a number of unique visitors.
The source data is rotating Apache access logs in the format:
192.114.71.13 - - [05/Jan/2014:19:10:19 +0100] "GET / HTTP/1.1" 302 186
"http:.../ref_if_transmitted.html" "Mozilla, browser version etc...."
Also, I prefer to avoid external services, free or otherwise, which do data
collection or simply isn't what I need, not to mention they can slow a site
down while connecting to their external servers unnecessarily sharing
visitor data to third-parties using cookies etc., which in all defeats the
general purpose of facilitating a positive user experience that should
ideally help build up traffic in the first place....
Any ideas, including home-grown open-source perl based logfile processing
solutions, would be most welcome!
Many thanks,
Tuxedo
p/page and p/IP in 24 hour cycles, so if several visits to a particular
page comes from a same IP in a 24 hours time frame, it counts as one
visitor to that page. While if the same visitor comes back to the same
page, say 25 hours later counting from his previous IP and page count, the
count can increment for that page again. In reality, this may represent a
repeat visitor or another visitor having been assigned the same IP.
Depending on type of web traffic and overall rotating IPs cycles by
different net connection service providers, this type of solution may at
best provide a highly approximate overview of a number of unique visitors.
The source data is rotating Apache access logs in the format:
192.114.71.13 - - [05/Jan/2014:19:10:19 +0100] "GET / HTTP/1.1" 302 186
"http:.../ref_if_transmitted.html" "Mozilla, browser version etc...."
Also, I prefer to avoid external services, free or otherwise, which do data
collection or simply isn't what I need, not to mention they can slow a site
down while connecting to their external servers unnecessarily sharing
visitor data to third-parties using cookies etc., which in all defeats the
general purpose of facilitating a positive user experience that should
ideally help build up traffic in the first place....
Any ideas, including home-grown open-source perl based logfile processing
solutions, would be most welcome!
Many thanks,
Tuxedo