2019年3月4日“robots.txt” contains 429 entries which should be manually viewed 正常情况下,robos.txt 文件是不会产生报错信息的,所以尝试另一种思路,进行网站的目录暴破,通常我们只会对后台管理页面感兴趣,但是那个会要求输入密码,所以只需要过滤 401响应代码的页面 ...
2024年11月29日15、XCTF Training-WWW-Robots 一打开网站就看到这行字In this little training challenge, you are going to learn about the Robots_exclusion_standard.The robots.txt file is used by web crawlers to check if they are allowed to crawl and index your website or only parts of it.Sometimes these fil...