Sha256: 734b7facc31e8a922e505f7e3b26cb6152afed5eb88a9f99eec3ed79eb5ea033
Contents?: true
Size: 1.3 KB
Versions: 6
Compression:
Stored size: 1.3 KB
Contents
# frozen_string_literal: true ## # This file is part of WhatWeb and may be subject to # redistribution and commercial restrictions. Please see the WhatWeb # web site for more information on licensing and terms of use. # http://www.morningstarsecurity.com/research/whatweb ## # Version 0.3 # 2011-03-23 # # Removed aggressive section ## # Version 0.2 # # Added aggressive `/robots.txt` retrieval ## WhatWeb::Plugin.define "robots_txt" do @author = "Brendan Coles <bcoles@gmail.com>" # 2010-10-22 @version = "0.3" @description = "This plugin identifies robots.txt files and extracts both allowed and disallowed directories. - More Info: http://www.robotstxt.org/" # Google results as at 2011-03-23 # # 920 for inurl:robots.txt filetype:txt # Passive # def passive(target) m = [] # Extract directories if current file is robots.txt if (target.uri.path.to_s == "/robots.txt") && target.body =~ /^User-agent:/i # File Exists m << { name: "File Exists" } # Disallow if /^Disallow:[\s]*(.+)$/i.match?(target.body) m << { string: target.body.scan(/^Disallow:[\s]*(.+)/i) } end # Allow if /^Allow:[\s]*(.+)$/i.match?(target.body) m << { string: target.body.scan(/^Allow:[\s]*(.+)/i) } end end # Return passive matches m end end
Version data entries
6 entries across 6 versions & 1 rubygems