Webbot activity on May 23, 2025

name activity Country
IP 216.73.216.30 1 page UNITED STATES
SemrushBotcrawler1 page UNITED STATES
IP 216.73.216.182 2 pages in  10s UNITED STATES
IP 216.73.216.158 1 page UNITED STATES
IP 216.73.216.194 2 pages in  3s UNITED STATES
IP 194.67.210.77  Spammer1 page RUSSIAN FEDERATION
IP 185.5.249.185  Spammer1 page RUSSIAN FEDERATION
IP 194.67.207.94  Spammer1 page RUSSIAN FEDERATION
IP 185.125.219.13  Spammer1 page RUSSIAN FEDERATION
IP 194.67.207.9  Spammer1 page RUSSIAN FEDERATION
IP 193.124.191.92  Spammer1 page RUSSIAN FEDERATION
IP 94.142.141.230  Spammer1 page RUSSIAN FEDERATION
IP 185.105.116.12  Spammer1 page RUSSIAN FEDERATION
IP 194.67.207.55  Spammer1 page RUSSIAN FEDERATION
IP 185.125.219.59  Spammer1 page RUSSIAN FEDERATION
IP 105.214.79.73  Spammer6 pages in  22s SOUTH AFRICA
IP 216.73.216.9 1 page UNITED STATES
IP 146.185.135.37  Spammer6 pages in  10s NETHERLANDS
IP 192.111.134.10  Spammer6 pages in  11s UNITED STATES
IP 216.73.216.44 1 page UNITED STATES
511,587 visits of identified bots
about 85 a day in 2025
185 today at 22:27 (+1 visitors)
zombies : 13 visits / 14 requests - spammers : 45 visits

Where does the data come from?

It has been extracted from $_SERVER["HTTP_USER_AGENT"] -$HTTP_USER_AGENT with PHP 3-, $_SERVER ["QUERY_STRING"], $_SERVER ["SCRIPT_NAME"] and gethostbyaddr().
As this website host (free.fr) sometimes filters access, it is biased.
The site statistics do not consider robots as visitors. The browser and country they show are ignored.
Even if they read a few pages they are stored once except if they come back after more than 10 minutes (for Google Desktop or semrush after more than 30 minutes).
This list gives approximate information as it assumes a perfect connection to MySQL which is not the case for this site. But web hosting here is free, so...

Robots represent at least 78 % of the visits in 2025:

Robot Detection

This routine is commented in the page about webbot traps.
Logging other visitors' user agent is necessary to update the lists of webbots visiting the site and of their User Agents.

The Data Table

Here is the structure of the table robots I use:
#
# Structure of the table `robots`
#
CREATE TABLE `robots` (
  `timeoflastaccess` int(10) unsigned NOT NULL default '0',
  `timeofarrival` int(10) unsigned NOT NULL default '0',
  `nameofbot` varchar(64) NOT NULL default '',
  `lastpage` varchar(30) NOT NULL default '',
  `numberofpages` mediumint(8) unsigned NOT NULL default '0',
  KEY `timeoflastaccess` (`timeoflastaccess`),
  KEY `timeofarrival` (`timeofarrival`),
  KEY `nameofbot` (`nameofbot`),
  KEY `numberofpages` (`numberofpages`)
) TYPE=MyISAM;

You can use double or datetime for times, double or int for numberofpages, if necessary, increase the number of caracters for lastpage.

Table Update

Data Display

We have the name of the robot, the time of its arrival, the time of the last page it loaded and the total number of pages indexed.
If there is one page and different values for timeoflastaccess and timeofarrival, then the page was reloaded.
I chose to display the number of pages loaded and the length of reading time.

A similar script is now online here

topTop of the page

With javascript

W3C XHTML 1.0
W3C CSS