2019-05-02 18:17:27 -04:00
|
|
|
# frozen_string_literal: true
|
|
|
|
|
2014-02-20 16:07:02 -05:00
|
|
|
module CrawlerDetection
|
2020-05-14 07:10:07 -04:00
|
|
|
WAYBACK_MACHINE_URL = "archive.org"
|
2017-09-28 22:31:50 -04:00
|
|
|
|
2018-01-16 00:28:11 -05:00
|
|
|
def self.to_matcher(string, type: nil)
|
2017-09-28 22:31:50 -04:00
|
|
|
escaped = string.split('|').map { |agent| Regexp.escape(agent) }.join('|')
|
2018-01-16 00:28:11 -05:00
|
|
|
|
|
|
|
if type == :real && Rails.env == "test"
|
|
|
|
# we need this bypass so we properly render views
|
|
|
|
escaped << "|Rails Testing"
|
|
|
|
end
|
|
|
|
|
2018-01-15 23:41:13 -05:00
|
|
|
Regexp.new(escaped, Regexp::IGNORECASE)
|
2017-09-28 22:31:50 -04:00
|
|
|
end
|
2015-02-14 09:24:51 -05:00
|
|
|
|
2019-06-02 22:13:32 -04:00
|
|
|
def self.crawler?(user_agent, via_header = nil)
|
2020-05-14 07:10:07 -04:00
|
|
|
return true if user_agent.nil? || user_agent&.include?(WAYBACK_MACHINE_URL) || via_header&.include?(WAYBACK_MACHINE_URL)
|
2018-01-16 00:28:11 -05:00
|
|
|
|
2017-09-28 22:31:50 -04:00
|
|
|
# this is done to avoid regenerating regexes
|
2018-01-15 23:41:13 -05:00
|
|
|
@non_crawler_matchers ||= {}
|
2017-09-28 22:31:50 -04:00
|
|
|
@matchers ||= {}
|
2018-01-15 23:41:13 -05:00
|
|
|
|
2018-01-16 00:28:11 -05:00
|
|
|
possibly_real = (@non_crawler_matchers[SiteSetting.non_crawler_user_agents] ||= to_matcher(SiteSetting.non_crawler_user_agents, type: :real))
|
2018-01-15 23:41:13 -05:00
|
|
|
|
|
|
|
if user_agent.match?(possibly_real)
|
|
|
|
known_bots = (@matchers[SiteSetting.crawler_user_agents] ||= to_matcher(SiteSetting.crawler_user_agents))
|
2018-06-20 20:56:46 -04:00
|
|
|
if user_agent.match?(known_bots)
|
|
|
|
bypass = (@matchers[SiteSetting.crawler_check_bypass_agents] ||= to_matcher(SiteSetting.crawler_check_bypass_agents))
|
|
|
|
!user_agent.match?(bypass)
|
|
|
|
else
|
|
|
|
false
|
|
|
|
end
|
2018-01-15 23:41:13 -05:00
|
|
|
else
|
|
|
|
true
|
|
|
|
end
|
|
|
|
|
2014-02-14 17:10:08 -05:00
|
|
|
end
|
2018-03-15 17:10:45 -04:00
|
|
|
|
|
|
|
# Given a user_agent that returns true from crawler?, should its request be allowed?
|
|
|
|
def self.allow_crawler?(user_agent)
|
2020-07-26 20:23:54 -04:00
|
|
|
return true if SiteSetting.allowed_crawler_user_agents.blank? &&
|
|
|
|
SiteSetting.blocked_crawler_user_agents.blank?
|
2018-03-15 17:10:45 -04:00
|
|
|
|
2020-07-26 20:23:54 -04:00
|
|
|
@allowlisted_matchers ||= {}
|
|
|
|
@blocklisted_matchers ||= {}
|
2018-03-15 17:10:45 -04:00
|
|
|
|
2020-07-26 20:23:54 -04:00
|
|
|
if SiteSetting.allowed_crawler_user_agents.present?
|
|
|
|
allowlisted = @allowlisted_matchers[SiteSetting.allowed_crawler_user_agents] ||= to_matcher(SiteSetting.allowed_crawler_user_agents)
|
|
|
|
!user_agent.nil? && user_agent.match?(allowlisted)
|
2018-03-15 17:10:45 -04:00
|
|
|
else
|
2020-07-26 20:23:54 -04:00
|
|
|
blocklisted = @blocklisted_matchers[SiteSetting.blocked_crawler_user_agents] ||= to_matcher(SiteSetting.blocked_crawler_user_agents)
|
|
|
|
user_agent.nil? || !user_agent.match?(blocklisted)
|
2018-03-15 17:10:45 -04:00
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def self.is_blocked_crawler?(user_agent)
|
|
|
|
crawler?(user_agent) && !allow_crawler?(user_agent)
|
|
|
|
end
|
2014-02-14 17:10:08 -05:00
|
|
|
end
|