Webbot activity on May 23, 2025
name | activity | Country | ||
IP 216.73.216.30 | 1 page | UNITED STATES | ||
SemrushBot | crawler | 1 page | UNITED STATES | |
IP 216.73.216.182 | 2 pages in 10s | UNITED STATES | ||
IP 216.73.216.158 | 1 page | UNITED STATES | ||
IP 216.73.216.194 | 2 pages in 3s | UNITED STATES | ||
IP 194.67.210.77 | Spammer | 1 page | RUSSIAN FEDERATION | |
IP 185.5.249.185 | Spammer | 1 page | RUSSIAN FEDERATION | |
IP 194.67.207.94 | Spammer | 1 page | RUSSIAN FEDERATION | |
IP 185.125.219.13 | Spammer | 1 page | RUSSIAN FEDERATION | |
IP 194.67.207.9 | Spammer | 1 page | RUSSIAN FEDERATION | |
IP 193.124.191.92 | Spammer | 1 page | RUSSIAN FEDERATION | |
IP 94.142.141.230 | Spammer | 1 page | RUSSIAN FEDERATION | |
IP 185.105.116.12 | Spammer | 1 page | RUSSIAN FEDERATION | |
IP 194.67.207.55 | Spammer | 1 page | RUSSIAN FEDERATION | |
IP 185.125.219.59 | Spammer | 1 page | RUSSIAN FEDERATION | |
IP 105.214.79.73 | Spammer | 6 pages in 22s | SOUTH AFRICA | |
IP 216.73.216.9 | 1 page | UNITED STATES | ||
IP 146.185.135.37 | Spammer | 6 pages in 10s | NETHERLANDS | |
IP 192.111.134.10 | Spammer | 6 pages in 11s | UNITED STATES | |
IP 216.73.216.44 | 1 page | UNITED STATES | ||
511,587 visits of identified bots about 85 a day in 2025 185 today at 22:27 (+1 visitors) zombies : 13 visits / 14 requests - spammers : 45 visits |
Where does the data come from?
It has been extracted from $_SERVER["HTTP_USER_AGENT"] -$HTTP_USER_AGENT with PHP 3-, $_SERVER ["QUERY_STRING"], $_SERVER ["SCRIPT_NAME"] and gethostbyaddr().As this website host (free.fr) sometimes filters access, it is biased.
The site statistics do not consider robots as visitors. The browser and country they show are ignored.
Even if they read a few pages they are stored once except if they come back after more than 10 minutes (for Google Desktop or semrush after more than 30 minutes).
This list gives approximate information as it assumes a perfect connection to MySQL which is not the case for this site. But web hosting here is free, so...
Robots represent at least 78 % of the visits in 2025:
- 30,97 % of these visits come from search engine bots, site indexing, analysis, link or code checking, SEO optimisation...
- 18,28 % come from spammers (60,13 % try to send emails, the 39,87 % left spam the log files).
- 50,75 % are looking for the site vulnerabilities.
Robot Detection
This routine is commented in the page about webbot traps.
Logging other visitors' user agent is necessary to update the lists of webbots visiting the site and of their User Agents.
The Data Table
Here is the structure of the table robots I use:# # Structure of the table `robots` # CREATE TABLE `robots` ( `timeoflastaccess` int(10) unsigned NOT NULL default '0', `timeofarrival` int(10) unsigned NOT NULL default '0', `nameofbot` varchar(64) NOT NULL default '', `lastpage` varchar(30) NOT NULL default '', `numberofpages` mediumint(8) unsigned NOT NULL default '0', KEY `timeoflastaccess` (`timeoflastaccess`), KEY `timeofarrival` (`timeofarrival`), KEY `nameofbot` (`nameofbot`), KEY `numberofpages` (`numberofpages`) ) TYPE=MyISAM;
You can use double or datetime for times, double or int for numberofpages, if necessary, increase the number of caracters for lastpage.
Table Update
Data Display
We have the name of the robot, the time of its arrival, the time of the last page it loaded and the total number of pages indexed.
If there is one page and different values for timeoflastaccess and timeofarrival, then the page was reloaded.
I chose to display the number of pages loaded and the length of reading time.
A similar script is now online here