98 lines
3.3 KiB
Ruby
98 lines
3.3 KiB
Ruby
require 'open-uri'
|
|
require 'nokogiri'
|
|
require 'excon'
|
|
require 'final_destination'
|
|
|
|
module Jobs
|
|
class CrawlTopicLink < Jobs::Base
|
|
|
|
class ReadEnough < StandardError; end
|
|
|
|
def self.max_chunk_size(uri)
|
|
# Amazon leaves the title until very late. Normally it's a bad idea to make an exception for
|
|
# one host but amazon is a big one.
|
|
return 80 if uri.host =~ /amazon\.(com|ca|co\.uk|es|fr|de|it|com\.au|com\.br|cn|in|co\.jp|com\.mx)$/
|
|
|
|
# Default is 10k
|
|
10
|
|
end
|
|
|
|
# Fetch the beginning of a HTML document at a url
|
|
def self.fetch_beginning(url)
|
|
# Never crawl in test mode
|
|
return if Rails.env.test?
|
|
|
|
fd = FinalDestination.new(url)
|
|
uri = fd.resolve
|
|
return "" unless uri
|
|
|
|
result = ""
|
|
streamer = lambda do |chunk, _, _|
|
|
result << chunk
|
|
|
|
# Using exceptions for flow control is really bad, but there really seems to
|
|
# be no sane way to get a stream to stop reading in Excon (or Net::HTTP for
|
|
# that matter!)
|
|
raise ReadEnough.new if result.size > (CrawlTopicLink.max_chunk_size(uri) * 1024)
|
|
end
|
|
Excon.get(uri.to_s, response_block: streamer, read_timeout: 20, headers: fd.request_headers)
|
|
result
|
|
|
|
rescue Excon::Errors::SocketError => ex
|
|
return result if ex.socket_error.is_a?(ReadEnough)
|
|
raise
|
|
rescue ReadEnough
|
|
result
|
|
end
|
|
|
|
def execute(args)
|
|
raise Discourse::InvalidParameters.new(:topic_link_id) unless args[:topic_link_id].present?
|
|
|
|
topic_link = TopicLink.find_by(id: args[:topic_link_id], internal: false, crawled_at: nil)
|
|
return if topic_link.blank?
|
|
|
|
# Look for a topic embed for the URL. If it exists, use its title and don't crawl
|
|
topic_embed = TopicEmbed.where(embed_url: topic_link.url).includes(:topic).references(:topic).first
|
|
# topic could be deleted, so skip
|
|
if topic_embed && topic_embed.topic
|
|
TopicLink.where(id: topic_link.id).update_all(['title = ?, crawled_at = CURRENT_TIMESTAMP', topic_embed.topic.title[0..255]])
|
|
return
|
|
end
|
|
|
|
begin
|
|
crawled = false
|
|
|
|
# Special case: Images
|
|
# If the link is to an image, put the filename as the title
|
|
if FileHelper.is_image?(topic_link.url)
|
|
uri = URI(topic_link.url)
|
|
filename = File.basename(uri.path)
|
|
crawled = (TopicLink.where(id: topic_link.id).update_all(["title = ?, crawled_at = CURRENT_TIMESTAMP", filename]) == 1)
|
|
end
|
|
|
|
unless crawled
|
|
# Fetch the beginning of the document to find the title
|
|
result = CrawlTopicLink.fetch_beginning(topic_link.url)
|
|
doc = Nokogiri::HTML(result)
|
|
if doc
|
|
title = doc.at('title').try(:inner_text)
|
|
if title.present?
|
|
title.gsub!(/\n/, ' ')
|
|
title.gsub!(/ +/, ' ')
|
|
title.strip!
|
|
if title.present?
|
|
crawled = (TopicLink.where(id: topic_link.id).update_all(['title = ?, crawled_at = CURRENT_TIMESTAMP', title[0..254]]) == 1)
|
|
end
|
|
end
|
|
end
|
|
end
|
|
rescue Exception
|
|
# If there was a connection error, do nothing
|
|
ensure
|
|
TopicLink.where(id: topic_link.id).update_all('crawled_at = CURRENT_TIMESTAMP') if !crawled && topic_link.present?
|
|
end
|
|
end
|
|
|
|
end
|
|
end
|